What is: S-shaped ReLU?
Source | Deep Learning with S-shaped Rectified Linear Activation Units |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
The S-shaped Rectified Linear Unit, or SReLU, is an activation function for neural networks. It learns both convex and non-convex functions, imitating the multiple function forms given by the two fundamental laws, namely the Webner-Fechner law and the Stevens law, in psychophysics and neural sciences. Specifically, SReLU consists of three piecewise linear functions, which are formulated by four learnable parameters.
The SReLU is defined as a mapping:
where , and are learnable parameters of the network and indicates that the SReLU can differ in different channels. The parameter represents the slope of the right line with input above a set threshold. and are thresholds in positive and negative directions respectively.
Source: Activation Functions