Loading Now

Summary of Population Transformer: Learning Population-level Representations Of Neural Activity, by Geeling Chau et al.


Population Transformer: Learning Population-level Representations of Neural Activity

by Geeling Chau, Christopher Wang, Sabera Talukder, Vighnesh Subramaniam, Saraswati Soedarmadji, Yisong Yue, Boris Katz, Andrei Barbu

First submitted to arxiv on: 5 Jun 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Neurons and Cognition (q-bio.NC)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
We propose a self-supervised framework, called the Population Transformer (PopT), that learns population-level codes for arbitrary ensembles of neural recordings at scale. This approach addresses key challenges in scaling models with neural time-series data by enabling learned aggregation of multiple spatially-sparse data channels. The PopT stacks on top of pretrained temporal embeddings and enhances downstream decoding by lowering the amount of data required while increasing accuracy, even on held-out subjects and tasks. Compared to end-to-end methods, this approach is computationally lightweight, achieving similar or better decoding performance. We demonstrate the framework’s generalizability to multiple time-series embeddings and neural data modalities. Our results show that PopT can be used not only for decoding but also for extracting neuroscience insights from large amounts of data.
Low GrooveSquid.com (original content) Low Difficulty Summary
We developed a new way to analyze brain signals recorded from many people at once. This method, called the Population Transformer (PopT), helps us understand how different brain areas work together by combining information from many electrodes. PopT is special because it can learn from a small amount of data and then be used for other tasks, like predicting what someone will do next. Our approach is faster and more accurate than some existing methods. We also showed that PopT can help us discover new things about the brain just by looking at its patterns.

Keywords

» Artificial intelligence  » Self supervised  » Time series  » Transformer