Variational autoencoder (VAE)

Variational autoencoder

A variational autoencoder (VAE) is a type of artificial neural network architecture used in machine learning. It is a probabilistic generative model that consists of an encoder and decoder neural network. The encoder maps the input to a latent space, while the decoder maps from the latent space back to the input space to generate data points. VAEs can be used for unsupervised, semi-supervised, and supervised learning tasks.

3 courses cover this concept

11-785 Introduction to Deep Learning

Carnegie Mellon University

Spring 2020

This course provides a comprehensive introduction to deep learning, starting from foundational concepts and moving towards complex topics such as sequence-to-sequence models. Students gain hands-on experience with PyTorch and can fine-tune models through practical assignments. A basic understanding of calculus, linear algebra, and Python programming is required.

No concepts data

+ 40 more concepts

CS231n: Deep Learning for Computer Vision

Stanford University

Spring 2022

This is a deep-dive into the details of deep learning architectures for visual recognition tasks. The course provides students with the ability to implement, train their own neural networks and understand state-of-the-art computer vision research. It requires Python proficiency and familiarity with calculus, linear algebra, probability, and statistics.

No concepts data

+ 55 more concepts

CSCI 1470/2470 Deep Learning

Brown University

Spring 2022

Brown University's Deep Learning course acquaints students with the transformative capabilities of deep neural networks in computer vision, NLP, and reinforcement learning. Using the TensorFlow framework, topics like CNNs, RNNs, deepfakes, and reinforcement learning are addressed, with an emphasis on ethical applications and potential societal impacts.

No concepts data

+ 40 more concepts