BERTBase
BERTbase is the base configuration of BERT (Bidirectional Encoder Representations from Transformers), a language representation model introduced by Google AI Language in 2018. It uses a bidirectional Transformer encoder to learn contextualized word representations from unlabeled text, enabling transfer learning for a wide range of natural language processing tasks. BERTbase is one of the two primary model sizes released with BERT, the other being BERT-large.
Model architecture and parameters: BERTbase consists of 12 Transformer encoder layers, a hidden size of 768,
Pretraining: The model is pretrained on large unlabeled text using two objectives: Masked Language Modeling (MLM),
Fine-tuning and use: After pretraining, BERTbase can be fine-tuned with task-specific layers on downstream NLP tasks
Comparison to BERT-large: The base model is smaller and faster to run than BERT-large, which uses 24