TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/GCT

GCT

Gated Channel Transformation

GeneralIntroduced 20008 papers
Source Paper

Description

GCT first collects global information by computing the l2-norm of each channel. Next, a learnable vector α\alphaα is applied to scale the feature. Then a competition mechanism is adopted by channel normalization to interact between channels.

Unlike previous methods, GCT first collects global information by computing the l2l_{2}l2​-norm of each channel. Next, a learnable vector α\alphaα is applied to scale the feature. Then a competition mechanism is adopted by channel normalization to interact between channels. Like other common normalization methods, a learnable scale parameter γ\gammaγ and bias β\betaβ are applied to rescale the normalization. However, unlike previous methods, GCT adopts tanh activation to control the attention vector. Finally, it not only multiplies the input by the attention vector but also adds an identity connection. GCT can be written as: \begin{align} s = F_\text{gct}(X, \theta) & = \tanh (\gamma CN(\alpha \text{Norm}(X)) + \beta) \end{align} \begin{align} Y & = s X + X \end{align}

where α\alphaα, β\betaβ and γ\gammaγ are trainable parameters. Norm(⋅)\text{Norm}(\cdot)Norm(⋅) indicates the L2L2L2-norm of each channel. CNCNCN is channel normalization.

A GCT block has fewer parameters than an SE block, and as it is lightweight, can be added after each convolutional layer of a CNN.

Papers Using This Method

STEAM: Squeeze and Transform Enhanced Attention Module2024-12-12Generative causal testing to bridge data-driven models and scientific theories in language neuroscience2024-10-01TelTrans: Applying Multi-Type Telecom Data to Transportation Evaluation and Prediction via Multifaceted Graph Modeling2024-01-06Granger Causality for Predictability in Dynamic Mode Decomposition2022-10-23Generalised Co-Salient Object Detection2022-08-20GCT: Graph Co-Training for Semi-Supervised Few-Shot Learning2022-03-15Integrating Fréchet distance and AI reveals the evolutionary trajectory and origin of SARS-CoV-22021-10-14Gated Channel Transformation for Visual Recognition2019-09-25