# Variational Autoencoder (VAE)

YouTube search... ...Google search

- Neural Network Zoo | Fjodor Van Veen
- Autoencoder (AE) / Encoder-Decoder
- Clustering
- Generative Tensorial Reinforcement Learning (GENTRL)
- Deep Belief Network (DBN)
- Restricted Boltzmann Machine (RBM)
- Digit Fantasies by a Deep Generative Model | Durk Kingma
- Morphing Faces | Vincent Dumolin

Variational autoencoders (VAE) have the same architecture as AEs but are “taught” something else: an approximated probability distribution of the input samples. It’s a bit back to the roots as they are bit more closely related to BMs and RBMs. They do however rely on Bayesian mathematics regarding probabilistic inference and independence, as well as a re-parametrisation trick to achieve this different representation. The inference and independence parts make sense intuitively, but they rely on somewhat complex mathematics. The basics come down to this: take influence into account. If one thing happens in one place and something else happens somewhere else, they are not necessarily related. If they are not related, then the error propagation should consider that. This is a useful approach because neural networks are large graphs (in a way), so it helps if you can rule out influence from some nodes to other nodes as you dive into deeper layers. Kingma, Diederik P., and Max Welling. “Auto-encoding variational bayes.” arXiv preprint arXiv:1312.6114 (2013).