SentenceBERT
Sentence-BERT is a modification of the BERT architecture designed to produce semantically meaningful sentence embeddings for efficient sentence-level similarity tasks. It adapts BERT through Siamese and triplet network setups to fine-tune the model so that embeddings of similar sentences are close in a fixed-dimensional space when measured by cosine similarity or related metrics. The approach relies on a pooling strategy to convert token representations into fixed-size sentence vectors, commonly using mean-pooling over the last hidden state, though other pooling methods are also used.
During training, Sentence-BERT is fine-tuned on sentence pairs drawn from natural language inference and paraphrase datasets,
Sentence-BERT was introduced by Nils Reimers and Iryna Gurevych in 2019, building on the capabilities of BERT