KreuzentropieLoss
KreuzentropieLoss, commonly referred to in English as cross-entropy loss, is a widely used loss function in supervised learning for probabilistic classification tasks. It measures the distance between a model’s predicted probability distribution over classes and the true distribution, typically represented as a one-hot vector for the target class. The function is derived from maximum likelihood estimation and is closely related to the Kullback–Leibler divergence.
For a single example with true distribution y and predicted distribution p, the loss is L = -
Backpropagation yields simple gradients, with respect to z, proportional to p - y, which facilitates learning.
Variants and implementations: Binary cross-entropy for binary classification; categorical cross-entropy for multi-class single-label; sparse categorical cross-entropy
Applications and limitations: It is standard in neural networks for image, text, and general classification tasks.