TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/ReLIC

ReLIC

GeneralIntroduced 200012 papers
Source Paper

Description

ReLIC, or Representation Learning via Invariant Causal Mechanisms, is a self-supervised learning objective that enforces invariant prediction of proxy targets across augmentations through an invariance regularizer which yields improved generalization guarantees.

We can write the objective as:

\underset{X}{\mathbb{E}} \underset{\sim\_{l k}, a\_{q \mathcal{A}}}{\mathbb{E}} \sum_{b \in\left\(a\_{l k}, a\_{q t}\right\)} \mathcal{L}\_{b}\left(Y^{R}, f(X)\right) \text { s.t. } K L\left(p^{d o\left(a\_{l k}\right)}\left(Y^{R} \mid f(X)\right), p^{d o\left(a\_{q t}\right)}\left(Y^{R} \mid f(X)\right)\right) \leq \rho

where L\mathcal{L}L is the proxy task loss and KLK LKL is the Kullback-Leibler (KL) divergence. Note that any distance measure on distributions can be used in place of the KL divergence.

Concretely, as proxy task we associate to every datapoint x_ix\_{i}x_i the label y_iR=iy\_{i}^{R}=iy_iR=i. This corresponds to the instance discrimination task, commonly used in contrastive learning. We take pairs of points (x_i,x_j)\left(x\_{i}, x\_{j}\right)(x_i,x_j) to compute similarity scores and use pairs of augmentations a_lk=(a_l,a_k)∈a\_{l k}=\left(a\_{l}, a\_{k}\right) \ina_lk=(a_l,a_k)∈ A×A\mathcal{A} \times \mathcal{A}A×A to perform a style intervention. Given a batch of samples \left\(x\_{i}\right\)\_{i=1}^{N} \sim \mathcal{D}, we use

pdo(a_lk)(YR=j∣f(x_i))∝exp⁡(ϕ(f(x_ia_l),h(x_ja_k))/τ)p^{d o\left(a\_{l k}\right)}\left(Y^{R}=j \mid f\left(x\_{i}\right)\right) \propto \exp \left(\phi\left(f\left(x\_{i}^{a\_{l}}\right), h\left(x\_{j}^{a\_{k}}\right)\right) / \tau\right)pdo(a_lk)(YR=j∣f(x_i))∝exp(ϕ(f(x_ia_l),h(x_ja_k))/τ)

with xax^{a}xa data augmented with aaa and τ\tauτ a softmax temperature parameter. We encode fff using a neural network and choose hhh to be related to fff, e.g. h=fh=fh=f or as a network with an exponential moving average of the weights of fff (e.g. target networks). To compare representations we use the function ϕ(f(x_i),h(x_j))=⟨g(f(x_i)),g(h(x_j))⟩\phi\left(f\left(x\_{i}\right), h\left(x\_{j}\right)\right)=\left\langle g\left(f\left(x\_{i}\right)\right), g\left(h\left(x\_{j}\right)\right)\right\rangleϕ(f(x_i),h(x_j))=⟨g(f(x_i)),g(h(x_j))⟩ where ggg is a fully-connected neural network often called the critic.

Combining these pieces, we learn representations by minimizing the following objective over the full set of data x_i∈Dx\_{i} \in \mathcal{D}x_i∈D and augmentations alk∈A×Aa_{l k} \in \mathcal{A} \times \mathcal{A}alk​∈A×A

−∑i=1N∑_a_lklog⁡exp⁡(ϕ(f(x_ial),h(x_ia_k))/τ)∑_m=1Mexp⁡(ϕ(f(x_ia_l),h(x_ma_k))/τ)+α∑_a_lk,a_qtKL(pdo(a_lk),pdo(a_qt))-\sum_{i=1}^{N} \sum\_{a\_{l k}} \log \frac{\exp \left(\phi\left(f\left(x\_{i}^{a_{l}}\right), h\left(x\_{i}^{a\_{k}}\right)\right) / \tau\right)}{\sum\_{m=1}^{M} \exp \left(\phi\left(f\left(x\_{i}^{a\_{l}}\right), h\left(x\_{m}^{a\_{k}}\right)\right) / \tau\right)}+\alpha \sum\_{a\_{l k}, a\_{q t}} K L\left(p^{d o\left(a\_{l k}\right)}, p^{d o\left(a\_{q t}\right)}\right)−i=1∑N​∑_a_lklog∑_m=1Mexp(ϕ(f(x_ia_l),h(x_ma_k))/τ)exp(ϕ(f(x_ial​),h(x_ia_k))/τ)​+α∑_a_lk,a_qtKL(pdo(a_lk),pdo(a_qt))

with MMM the number of points we use to construct the contrast set and α\alphaα the weighting of the invariance penalty. The shorthand pdo(a)p^{d o(a)}pdo(a) is used for pdo(a)(YR=j∣f(x_i))p^{d o(a)}\left(Y^{R}=j \mid f\left(x\_{i}\right)\right)pdo(a)(YR=j∣f(x_i)). The Figure shows a schematic of the RELIC objective.

Papers Using This Method

RELIC: Evaluating Compositional Instruction Following via Language Recognition2025-06-05A Collaborative Jade Recognition System for Mobile Devices Based on Lightweight and Large Models2025-02-20An RNA condensate model for the origin of life2024-12-06MRIFE: A Mask-Recovering and Interactive-Feature-Enhancing Semantic Segmentation Network For Relic Landslide Detection2024-11-26Machine-Learning Analysis of Radiative Decays to Dark Matter at the LHC2024-10-17ReLIC: A Recipe for 64k Steps of In-Context Reinforcement Learning for Embodied AI2024-10-03Symbolic Regression for Beyond the Standard Model Physics2024-05-28Self-Supervised Learning Through Efference Copies2022-10-17Mixed Anhydrides at the Intersection Between Peptide and RNA Autocatalytic Sets: Evolution of Biological Coding2022-05-25Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?2022-01-13CoBERL: Contrastive BERT for Reinforcement Learning2021-07-12Representation Learning via Invariant Causal Mechanisms2020-10-15