TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/AGLU

AGLU

Adaptive Generalised Linear Unit

GeneralIntroduced 20001 papers
Source Paper

Description

The Adaptive Parametric activation (APA) is defined as: APA(z,λ,κ)=(λexp(−κz)+1)1−λAPA(z,λ,κ) = (λ exp(−κz) + 1) ^{\frac{1}{−λ}}APA(z,λ,κ)=(λexp(−κz)+1)−λ1​, where λλλ and κκκ are learnable parameters. This activation function is a generalisation of the Sigmoid and the Gumbel activation functions and it is expressive and versatile. For example, APA can be used inside the channel attention mechanism instead of the Sigmoid activation, or it can be used inside the intermediate layers using the Adaptive Generalised Linear Unit (AGLU): AGLU(z,λ,κ)=zAPA(z,λ,κ)AGLU(z,λ,κ) = z APA(z,λ,κ)AGLU(z,λ,κ)=zAPA(z,λ,κ).

Papers Using This Method

Adaptive Parametric Activation2024-07-11