transformers

A neural network architecture based on self-attention mechanisms, introduced in "Attention is All You Need" (2017). The dominant paradigm for modern NLP and increasingly for vision tasks.