Gradient descent is an iterative optimization algorithm used to find the local minimum of a differentiable function. It is commonly used in machine learning for minimizing cost or loss functions and was first proposed by Augustin-Louis Cauchy in 1847. Stochastic gradient descent is a simple extension of gradient descent which is used to train most deep networks today.
Stanford University
Autumn 2022-2023
Stanford's CS 221 course teaches foundational principles and practical implementation of AI systems. It covers machine learning, game playing, constraint satisfaction, graphical models, and logic. A rigorous course requiring solid foundational skills in programming, math, and probability.
No concepts data
+ 88 more conceptsCarnegie Mellon University
Spring 2020
This course provides a comprehensive introduction to deep learning, starting from foundational concepts and moving towards complex topics such as sequence-to-sequence models. Students gain hands-on experience with PyTorch and can fine-tune models through practical assignments. A basic understanding of calculus, linear algebra, and Python programming is required.
No concepts data
+ 40 more conceptsCarnegie Mellon University
Spring 2018
A comprehensive exploration of machine learning theories and practical algorithms. Covers a broad spectrum of topics like decision tree learning, neural networks, statistical learning, and reinforcement learning. Encourages hands-on learning via programming assignments.
No concepts data
+ 55 more conceptsPrinceton University
Fall 2017
A thorough introduction to machine learning principles such as online learning, decision making, gradient-based learning, and empirical risk minimization. It also explores regression, classification, dimensionality reduction, ensemble methods, neural networks, and deep learning. The course material is self-contained and based on freely available resources.
No concepts data
+ 14 more concepts