ReLU
ReLU, short for rectified linear unit, is a widely used activation function in artificial neural networks. It is defined as f(x) = max(0, x), meaning it outputs x for positive inputs and 0 for negative inputs. The function is piecewise linear, and its derivative is 1 for x > 0 and 0 for x < 0; at x = 0 the derivative is not uniquely defined, but subgradients are typically used in training.
ReLU offers several practical advantages. It is computationally simple and fast to evaluate, which helps training
Despite its benefits, ReLU has limitations. Neurons can become “dead” if they consistently receive negative inputs,
Variants of ReLU address these issues. Leaky ReLU introduces a small slope for negative inputs; Parametric
ReLU remains a standard choice in many neural network designs and continues to be a baseline activation