TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/GLN

GLN

Gated Linear Network

GeneralIntroduced 20008 papers
Source Paper

Description

A Gated Linear Network, or GLN, is a type of backpropagation-free neural architecture. What distinguishes GLNs from contemporary neural networks is the distributed and local nature of their credit assignment mechanism; each neuron directly predicts the target, forgoing the ability to learn feature representations in favor of rapid online learning. Individual neurons can model nonlinear functions via the use of data-dependent gating in conjunction with online convex optimization.

GLNs are feedforward networks composed of many layers of gated geometric mixing neurons as shown in the Figure . Each neuron in a given layer outputs a gated geometric mixture of the predictions from the previous layer, with the final layer consisting of just a single neuron. In a supervised learning setting, a GLN\mathrm{GLN}GLN is trained on (side information, base predictions, label) triplets (z_t,p_t,x_t)t=1,2,3,…\left(z\_{t}, p\_{t}, x\_{t}\right)_{t=1,2,3, \ldots}(z_t,p_t,x_t)t=1,2,3,…​ derived from input-label pairs (z_t,x_t)\left(z\_{t}, x\_{t}\right)(z_t,x_t). There are two types of input to neurons in the network: the first is the side information z_tz\_{t}z_t, which can be thought of as the input features; the second is the input to the neuron, which will be the predictions output by the previous layer, or in the case of layer 0 , some (optionally) provided base predictions p_tp\_{t}p_t that typically will be a function of z_t.z\_{t} .z_t. Each neuron will also take in a constant bias prediction, which helps empirically and is essential for universality guarantees.

Weights are learnt in a Gated Linear Network using Online Gradient Descent (OGD) locally at each neuron. They key observation is that as each neuron (i,k)(i, k)(i,k) in layers i>0i>0i>0 is itself a gated geometric mixture, all of these neurons can be thought of as individually predicting the target. Given side information zzz , each neuron (i,k)(i, k)(i,k) suffers a loss convex in its active weights u:=w_ikc_ik(z)u:=w\_{i k c\_{i k}(z)}u:=w_ikc_ik(z) of

ℓ_t(u):=−log⁡(GEO⁡_u(xt;p_i−1))\ell\_{t}(u):=-\log \left(\operatorname{GEO}\_{u}\left(x_{t} ; p\_{i-1}\right)\right)ℓ_t(u):=−log(GEO_u(xt​;p_i−1))

Papers Using This Method

NoisyGL: A Comprehensive Benchmark for Graph Neural Networks under Label Noise2024-06-06SemiRetro: Semi-template framework boosts deep retrosynthesis prediction2022-02-12Zero-Shot Multi-View Indoor Localization via Graph Location Networks2020-08-06Gait Lateral Network: Learning Discriminative and Compact Representations for Gait Recognition2020-08-01A Generative Graph Method to Solve the Travelling Salesman Problem2020-07-09Gaussian Gated Linear Networks2020-06-10Online Learning in Contextual Bandits using Gated Linear Networks2020-02-21Gated Linear Networks2019-09-30