Probably Approximately Correct (PAC) learning is a framework for mathematical analysis of machine learning proposed by Leslie Valiant in 1984. It involves selecting a generalization function from a certain class of possible functions with the goal of having low generalization error with high probability. The model was later extended to include noise and introduced computational complexity theory concepts to machine learning.
Stanford University
Spring 2022
CS 168 provides a comprehensive introduction to modern algorithm concepts, covering hashing, dimension reduction, programming, gradient descent, and regression. It emphasizes both theoretical understanding and practical application, with each topic complemented by a mini-project. It's suitable for those who have taken CS107 and CS161.
No concepts data
+ 57 more concepts