Transformer
The Transformer is a deep learning architecture introduced by Vaswani et al. in the 2017 paper 'Attention Is All You Need,' which has become the dominant paradigm for large-scale language modeling and generative AI. It replaces sequential processing architectures like RNNs and LSTMs with a fully attention-based mechanism that processes all positions in a sequence simultaneously, enabling massive parallelization and more direct modeling of long-range dependencies.
The core innovation is multi-head self-attention: each token in a sequence attends to all other tokens through learned query, key, and value projections, allowing the model to dynamically weight the relevance of different contextual positions. This is combined with feed-forward networks, layer normalization, and residual connections to enable training at unprecedented scale — modern Transformer-based models contain hundreds of billions of parameters trained on trillions of tokens.
Transformers have demonstrated emergent capabilities at scale: few-shot learning, chain-of-thought reasoning, in-context learning, and analogical transfer appear in sufficiently large models despite not being explicitly trained for. These emergent properties are not universal to all neural architectures; they are specific to the Transformer design and its training regimen, suggesting that implementation details matter profoundly for which cognitive functions emerge.
The architecture is not limited to language. Vision Transformers (ViT) apply the same mechanism to image patches. Protein folding models like AlphaFold2 use attention to capture spatial relationships in molecular structures. The Transformer appears to be a general-purpose pattern-matching engine whose effectiveness depends on the availability of large, structured datasets and computational resources for training.
The Transformer is often described as a generic universal function approximator, but this description conceals a deeper truth: the specific mechanics of attention — the query-key-value decomposition, the softmax normalization, the residual pathways — are not arbitrary engineering choices. They are the precise conditions under which certain complex functions become learnable. Change the architecture, and the capabilities vanish. The Transformer is not a substrate-independent mind; it is a very specific implementation that happens to unlock certain cognitive functions at scale.