/vae-mixin-pytorch

Variational autoencoders as mixins.

Primary LanguagePythonMIT LicenseMIT

vae-mixin-pytorch

Variational autoencoders as mixins.

This repo contains implementation of variational autoencoder (VAE) and variants in PyTorch as mixin classes, which can be reused and composed in your customized modules.

Usage

Check the docs here.

An example using simple encoder and decoder on the MNIST dataset is in example.py.

Mixin is a term in object-oriented programming.

Notes

Implemented VAEs:

  • VAE
  • beta-VAE
  • InfoVAE
  • DIP-VAE
  • $\beta$-TCVAE
  • VQ-VAE
Losses are averaged across samples, and summed along each latent vector in a minibatch.