Temporal difference (TD) learning

Temporal difference learning

Temporal Difference (TD) learning is a model-free reinforcement learning method which uses bootstrapping to adjust predictions based on current estimates, rather than waiting for the final outcome. TD methods allow for predictions to be adjusted before the final outcome is known, allowing for more accurate predictions. TD methods are also related to the temporal difference model of animal learning.

2 courses cover this concept

CS 221 Artificial Intelligence: Principles and Techniques

Stanford University

Autumn 2022-2023

Stanford's CS 221 course teaches foundational principles and practical implementation of AI systems. It covers machine learning, game playing, constraint satisfaction, graphical models, and logic. A rigorous course requiring solid foundational skills in programming, math, and probability.

No concepts data

+ 88 more concepts

CS 294-40: Learning for robotics and control

UC Berkeley

Fall 2008

This advanced course focuses on the applications of machine learning in the robotics and control field. It covers a wide range of topics including Markov Decision Processes, control theories, estimation methodologies, and robotics principles. Recommended for graduate students.

No concepts data

+ 27 more concepts