Codistillation
Codistillation is a learning paradigm in machine learning in which multiple models are trained in concert and exchange information during training, typically by sharing softened output distributions rather than hard labels. In codistillation, each model acts as both student and teacher to the others, using the peers' predictions as auxiliary targets in addition to ground-truth labels. This approach is commonly used in semi-supervised learning, ensemble methods, and scenarios with limited labeled data.
Typically, training involves a combination of a standard supervised loss on labeled data and a distillation
Benefits include improved accuracy and calibration when labeled data are scarce, better utilization of unlabeled data,
See also: knowledge distillation, semi-supervised learning, co-training, ensemble learning.