TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/ELiSH

ELiSH

Exponential Linear Squashing Activation

GeneralIntroduced 20001 papers
Source Paper

Description

The Exponential Linear Squashing Activation Function, or ELiSH, is an activation function used for neural networks. It shares common properties with Swish, being made up of an ELU and a Sigmoid:

f(x)=x1+e−x if x≥0f\left(x\right) = \frac{x}{1+e^{-x}} \text{ if } x \geq 0 f(x)=1+e−xx​ if x≥0 f(x)=ex−11+e−x if x<0f\left(x\right) = \frac{e^{x} - 1}{1+e^{-x}} \text{ if } x < 0 f(x)=1+e−xex−1​ if x<0

The Sigmoid part of ELiSH improves information flow, while the linear parts solve issues of vanishing gradients.

Papers Using This Method

The Quest for the Golden Activation Function2018-08-02