TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/SReLU

SReLU

S-shaped ReLU

GeneralIntroduced 20002 papers
Source Paper

Description

The S-shaped Rectified Linear Unit, or SReLU, is an activation function for neural networks. It learns both convex and non-convex functions, imitating the multiple function forms given by the two fundamental laws, namely the Webner-Fechner law and the Stevens law, in psychophysics and neural sciences. Specifically, SReLU consists of three piecewise linear functions, which are formulated by four learnable parameters.

The SReLU is defined as a mapping:

f(x)=t_ir+ar_i(x_i−tr_i) if x_i≥tr_if\left(x\right) = t\_{i}^{r} + a^{r}\_{i}\left(x\_{i}-t^{r}\_{i}\right) \text{ if } x\_{i} \geq t^{r}\_{i}f(x)=t_ir+ar_i(x_i−tr_i) if x_i≥tr_i f(x)=x_i if tr_i>x>t_il f\left(x\right) = x\_{i} \text{ if } t^{r}\_{i} > x > t\_{i}^{l}f(x)=x_i if tr_i>x>t_il f(x)=t_il+al_i(x_i−tl_i) if x_i≤tl_if\left(x\right) = t\_{i}^{l} + a^{l}\_{i}\left(x\_{i}-t^{l}\_{i}\right) \text{ if } x\_{i} \leq t^{l}\_{i}f(x)=t_il+al_i(x_i−tl_i) if x_i≤tl_i

where tl_it^{l}\_{i}tl_i, tr_it^{r}\_{i}tr_i and al_ia^{l}\_{i}al_i are learnable parameters of the network iii and indicates that the SReLU can differ in different channels. The parameter ar_ia^{r}\_{i}ar_i represents the slope of the right line with input above a set threshold. tr_it^{r}\_{i}tr_i and tl_it^{l}\_{i}tl_i are thresholds in positive and negative directions respectively.

Source: Activation Functions

Papers Using This Method

Overcoming Overfitting and Large Weight Update Problem in Linear Rectifiers: Thresholded Exponential Rectified Linear Units2020-06-04Deep Learning with S-shaped Rectified Linear Activation Units2015-12-22