Relu2
Relu2, also known as the Swish-1 activation function, is a variant of the Rectified Linear Unit (ReLU) activation function used in artificial neural networks. It was introduced as part of the Swish family of activation functions, which aim to improve the performance of neural networks by providing a smooth and non-monotonic curve. The Swish-1 function is defined as:
f(x) = x * sigmoid(βx)
where β is a learnable parameter that controls the shape of the function. When β is set to
f(x) = x * sigmoid(x)
This variant is often referred to as Relu2 due to its similarity to the standard ReLU function,