selfsupervision
Self-supervision is a form of representation learning in which a model learns from unlabeled data by solving automatically generated tasks, called pretext tasks. The supervisory signal is derived from the data itself, not human annotations. The learned representations can be transferred to downstream tasks with limited labeled data, reducing labeling requirements compared to fully supervised learning.
Pretext tasks can be predictive, reconstructive, or contrastive. Predictive tasks require the model to infer missing
Notable self-supervised approaches in vision include colorization, inpainting, rotation prediction, and jigsaw puzzles; in natural language
Evaluation typically uses linear evaluation or fine-tuning on downstream tasks. Large unlabeled datasets are common, and
Applications include computer vision, natural language processing, speech, and time-series analysis. Self-supervision remains an active research