Temporal Difference (TD) learning is a model-free reinforcement learning method which uses bootstrapping to adjust predictions based on current estimates, rather than waiting for the final outcome. TD methods allow for predictions to be adjusted before the final outcome is known, allowing for more accurate predictions. TD methods are also related to the temporal difference model of animal learning.
Stanford University
Autumn 2022-2023
Stanford's CS 221 course teaches foundational principles and practical implementation of AI systems. It covers machine learning, game playing, constraint satisfaction, graphical models, and logic. A rigorous course requiring solid foundational skills in programming, math, and probability.
No concepts data
+ 88 more conceptsUC Berkeley
Fall 2008
This advanced course focuses on the applications of machine learning in the robotics and control field. It covers a wide range of topics including Markov Decision Processes, control theories, estimation methodologies, and robotics principles. Recommended for graduate students.
No concepts data
+ 27 more concepts