Shared latent-objective algebra #
We use this file to connect the latent generative models without forcing them into one artificial architecture. VAEs, VQ-VAEs, and GANs make different modeling choices, but their training objectives share a small amount of algebra:
- a base term, such as reconstruction loss or real-sample score regression;
- one or more regularizing/critic terms; and
- a scalar weight that controls the tradeoff.
Keeping that algebra in one place gives the model-specific files a common language. The VAE file can say "β-VAE is a weighted two-term objective"; the VQ-VAE file can say "VQ-VAE is a weighted three-term objective"; the GAN file can say "LSGAN is score-regression algebra over the same scalar loss vocabulary."
References:
- Kingma and Welling, "Auto-Encoding Variational Bayes", ICLR 2014.
- van den Oord, Vinyals, and Kavukcuoglu, "Neural Discrete Representation Learning", NeurIPS 2017.
- Mao et al., "Least Squares Generative Adversarial Networks", ICCV 2017.
A two-term latent objective: base + weight * regularizer.
- base : ℝ
Main data-fitting term, typically reconstruction or score regression.
- regularizer : ℝ
Latent regularizer, KL term, commitment loss, or critic penalty.
Instances For
Evaluate a weighted two-term objective.
Instances For
Evaluate a weighted three-term objective.
Instances For
A weighted two-term objective collapses to its base term when the regularizer is zero.
At weight zero, a weighted two-term objective ignores the regularizer.
If the regularizer is nonnegative, increasing the weight can only increase a two-term objective.
This is the algebraic core behind β-VAE monotonicity in the KL weight.
A weighted three-term objective collapses to base + middle when the regularizer is zero.
A weighted three-term objective collapses to base when both auxiliary terms are zero.
At weight zero, a weighted three-term objective keeps only its base and middle terms.
If the weighted regularizer is nonnegative, increasing the weight can only increase a three-term objective.
For VQ-VAE this is the commitment-weight monotonicity statement.