THE FUTURE IS HERE

Variational Autoencoders | Generative AI Animated

In this video you will learn everything about variational autoencoders. These generative models have been popular for more than a decade, and are still used in many applications.

If you want to dive even deeper into this topic, I would suggest you read the original paper from Kingma, and an overview he wrote later on:
– Auto-Encoding Variational Bayes https://arxiv.org/abs/1312.6114
– An Introduction to Variational Autoencoders https://arxiv.org/abs/1906.02691

If you want more accessible ressources, these blog posts by Matthew N. Bernstein are incredible to understand the different parts of the theory behind VAEs:
– Variational Autoencoders https://mbernste.github.io/posts/vae/
– Variational Inference https://mbernste.github.io/posts/variational_inference/
– The Evidence Lower Bound https://mbernste.github.io/posts/elbo/

Chapters:
00:00 Introduction
01:05 Context
06:20 General Principle of VAEs
08:53 Evidence Lower Bound
11:01 The Reparameterization Trick
14:05 Training and Inference
16:28 Limitations
18:40 Bonus: ELBO derivations

This video features animations created with Manim, inspired by Grant Sanderson’s work at @3blue1brown.
All the code for the animations of this video is available in the following github repository: https://github.com/ytdeepia/Variational-Autoencoders

If you enjoyed the content, please like, comment, and subscribe to support the channel!

#deeplearning #artificialintelligence #generativeai #machinelearning #manim #education #science