Transformervarianter
Transformervarianter is a broad term used in contemporary AI literature to describe the family of models built around the transformer architecture. It encompasses encoder-only, decoder-only, and encoder-decoder designs, as well as variants that modify attention patterns, training objectives, or modalities. The term is sometimes used in Nordic and European research contexts to refer to transformer variants as a collective category.
The concept arose from the original Transformer (Vaswani et al., 2017), which introduced self-attention and a
Core features across transformervarianter include multi-head self-attention, positional encoding, residual connections, and layer normalization. Variants differ
Common subtypes include encoder-only models for representation learning, decoder-only models for generation, and encoder-decoder models for
Applications span natural language processing, computer vision, multi-modal tasks, and research for code or biology. Challenges
See also: Transformer; BERT; GPT; T5; ViT; Longformer; Reformer.