Transformers and cortical waves: encoders for pulling in context across time

Bibliographic Collection: 
CARTA-Inspired Publication
Publication Type: Journal Article
Authors: Muller, Lyle; Churchland, Patricia S.; Sejnowski, Terrence J.
Year of Publication: 2024
Volume: 47
Issue: 10
Pagination: 788 - 802
Date Published: 2024/10/01/
Publication Language: eng
ISBN Number: 0166-2236
Keywords: Artificial intelligence, Brain, computational neuroscience, large language models, sensory processing, visual processing
Abstract:

The capabilities of transformer networks such as ChatGPT and other large language models (LLMs) have captured the world’s attention. The crucial computational mechanism underlying their performance relies on transforming a complete input sequence – for example, all the words in a sentence – into a long ‘encoding vector’ that allows transformers to learn long-range temporal dependencies in naturalistic sequences. Specifically, ‘self-attention’ applied to this encoding vector enhances temporal context in transformers by computing associations between pairs of words in the input sequence. We suggest that waves of neural activity traveling across single cortical areas, or multiple regions on the whole-brain scale, could implement a similar encoding principle. By encapsulating recent input history into a single spatial pattern at each moment in time, cortical waves may enable a temporal context to be extracted from sequences of sensory inputs, the same computational principle as that used in transformers.

DOI: https://doi.org/10.1016/j.tins.2024.08.006
Short Title: Trends in Neurosciences
Export: