Activation functions and their derivatives for a Neuron.
- Source:
Members
(static) identity
Simply passes the input to the output with no transformation.
Simply passes the input to the output with no transformation.
- Source:
(static) logistic
A smoothed step function or an 'S' shape.
A smoothed step function or an 'S' shape. Also called the sigmoid function, though there are many sigmoid functions.
- Source:
(static) optimalTanh
Modified hyperbolic tangent function.
Modified hyperbolic tangent function. Optimized for faster convergence.
- Source:
(static) rectifier
Simply max(0, x).
Simply max(0, x). Interestingly the derivative of the rectifier turns out to be the logistic function.
- Source:
(static) softplus
A smooth approximation of the rectifier.
A smooth approximation of the rectifier.
- Source:
(static) tanh
The hyperbolic tangent function.
The hyperbolic tangent function. A sigmoid curve, like the logistic function, except it has a range of (-1,+1). Often performs better than the logistic function because of its symmetry. Ideal for customization of multilayer perceptrons, particularly the hidden layers.
- Source: