encoderonly
Encoder-only models are transformer-based neural networks that focus on encoding input sequences into contextualized representations without producing autoregressive outputs. They are designed for understanding and representation tasks rather than text generation, distinguishing them from decoder-only and encoder-decoder architectures.
These models typically consist of multiple stacked transformer encoder layers that apply self-attention to the input
Common examples of encoder-only models include BERT, RoBERTa, ALBERT, ELECTRA, DistilBERT, and domain-specific variants like BioBERT.
Compared with decoder-only models, encoder-only architectures excel at understanding and embedding text rather than generating it.