IE5571: Reinforcement Learning and Dynamic Programming

4 Credits

This course introduces students to modeling and solution methods for problems in sequential decision-making. Applications include games such as Backgammon, Chess, and Go, as well as robotics, traffic control, resource management, and financial trading strategies. We begin with Markov Decision Processes and Dynamic Programming, and then move to model-free Reinforcement Learning (RL) approaches that include Monte Carlo and Temporal Difference learning. We then discuss extensions of these methods to problems of practical interest where it is necessary to employ a supervised learning method to obtain an optimal strategy or policy. Students will obtain hands-on experience by implementing RL methods in a modern programming language such as Python or Julia. Prereq: Knowledge of multivariable calculus, linear algebra, and probability at the undergraduate level.

View on University Catalog

All Instructors

B Average (2.917)Most Common: B (27%)

This total also includes data from semesters with unknown instructors.

15 students
SWFDCBA
  • 5.50

    /6

    Recommend
  • 5.50

    /6

    Effort
  • 6.00

    /6

    Understanding
  • 5.50

    /6

    Interesting
  • 5.50

    /6

    Activities


      Contribute on our Github

      Gopher Grades is maintained by Social Coding with data from Summer 2017 to Fall 2025 provided by the University in response to a public records request

      Not affiliated with the University of Minnesota

      Privacy Policy