TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Beta-VAE

Beta-VAE

Computer VisionIntroduced 200030 papers
Source Paper

Description

Beta-VAE is a type of variational autoencoder that seeks to discover disentangled latent factors. It modifies VAEs with an adjustable hyperparameter β\betaβ that balances latent channel capacity and independence constraints with reconstruction accuracy. The idea is to maximize the probability of generating the real data while keeping the distance between the real and estimated distributions small, under a threshold ϵ\epsilonϵ. We can use the Kuhn-Tucker conditions to write this as a single equation:

F(θ,ϕ,β;x,z)=E_q_ϕ(z∣x)[log⁡p_θ(x∣z)]−β[D_KL(log⁡q_θ(z∣x)∣∣p(z))−ϵ] \mathcal{F}\left(\theta, \phi, \beta; \mathbf{x}, \mathbf{z}\right) = \mathbb{E}\_{q\_{\phi}\left(\mathbf{z}|\mathbf{x}\right)}\left[\log{p}\_{\theta}\left(\mathbf{x}\mid\mathbf{z}\right)\right] - \beta\left[D\_{KL}\left(\log{q}\_{\theta}\left(\mathbf{z}\mid\mathbf{x}\right)||p\left(\mathbf{z}\right)\right) - \epsilon\right]F(θ,ϕ,β;x,z)=E_q_ϕ(z∣x)[logp_θ(x∣z)]−β[D_KL(logq_θ(z∣x)∣∣p(z))−ϵ]

where the KKT multiplier β\betaβ is the regularization coefficient that constrains the capacity of the latent channel z\mathbf{z}z and puts implicit independence pressure on the learnt posterior due to the isotropic nature of the Gaussian prior p(z)p\left(\mathbf{z}\right)p(z).

We write this again using the complementary slackness assumption to get the Beta-VAE formulation:

\mathcal{F}\left(\theta, \phi, \beta; \mathbf{x}, \mathbf{z}\right) \geq \mathcal{L}\left(\theta, \phi, \beta; \mathbf{x}, \mathbf{z}\right) = \mathbb{E}\_{q\_{\phi}\left(\mathbf{z}|\mathbf{x}\right)}\left[\log{p}\_{\theta}\left(\mathbf{x}\mid\mathbf{z}\right)\right] - \beta\{D}\_{KL}\left(\log{q}\_{\theta}\left(\mathbf{z}\mid\mathbf{x}\right)||p\left(\mathbf{z}\right)\right)

Papers Using This Method

Causal Intervention Framework for Variational Auto Encoder Mechanistic Interpretability2025-05-06Alternatives of Unsupervised Representations of Variables on the Latent Space2024-10-26Comparison of Autoencoder Encodings for ECG Representation in Downstream Prediction Tasks2024-10-03Lewis's Signaling Game as beta-VAE For Natural Word Lengths and Segments2023-11-08From Conceptual Spaces to Quantum Concepts: Formalising and Learning Structured Conceptual Models2023-11-06Learning minimal representations of stochastic processes with variational autoencoders2023-07-21Impact of Disentanglement on Pruning Neural Networks2023-07-19Identification of Rare Cortical Folding Patterns using Unsupervised Deep Learning2022-11-29Analysis of Master Vein Attacks on Finger Vein Recognition Systems2022-10-18TopicVAE: Topic-aware Disentanglement Representation Learning for Enhanced Recommendation2022-10-10The Conceptual VAE2022-03-21Designing Complex Experiments by Applying Unsupervised Machine Learning2021-09-29Be More Active! Understanding the Differences between Mean and Sampled Representations of Variational Autoencoders2021-09-26Finding simplicity: unsupervised discovery of features, patterns, and order parameters via shift-invariant variational autoencoders2021-06-23Recovering Barabási-Albert Parameters of Graphs through Disentanglement2021-05-03Learning Stable Representations with Full Encoder2021-03-25Dynamic $β$-VAEs for quantifying biodiversity by clustering optically recorded insect signals2021-02-10Autoencoding Slow Representations for Semi-supervised Data Efficient Regression2020-12-11AI Discovering a Coordinate System of Chemical Elements: Dual Representation by Variational Autoencoders2020-11-24Unsupervised anomaly localization using VAE and beta-VAE2020-05-19