Regularizers
Regularizers are techniques added to the loss function during training to discourage model complexity and prevent overfitting. By penalizing large or numerous weights, regularizers constrain the capacity of a model, encouraging simpler, more generalizable solutions.
Common forms include L1 regularization (lasso) and L2 regularization (ridge). L1 adds a penalty proportional to
Weight decay is a closely related concept often implemented as L2 regularization in optimization algorithms. Other
Data augmentation can act as a regularizer by increasing input diversity and robustness. Max-norm constraints limit
Regularizers are widely used across supervised learning tasks, including linear models, logistic regression, and neural networks.