Fall 2023 Decision, Risk & Operations Management B9120 section 001

(PhD) Dynamic Programming

Call Number 16932
Day & Time
Location
R 9:00am-12:15pm
430 Kravis Hall
Points 3
Grading Mode Standard
Approvals Required None
Instructor Daniel J Russo
Type LECTURE
Method of Instruction In-Person
Course Description

This course offers an advanced introduction Markov Decision Processes (MDPs)-a formalization of the problem of optimal sequential decision making under uncertainty-and Reinforcement Learning (RL)-a paradigm for learning from data to make near optimal sequential decisions. The first part of the course will cover foundational material on MDPs. We'll then look at the problem of estimating long run value from data, including popular RL algorithms like temporal difference learning and Q-learning. The final part of the course looks at the design and analysis of efficient exploration algorithms, i.e. those that intelligently probe the environment to collect data that improves decision quality. This a doctoral level course. Students should have experience with mathematical proofs, coding for numerical computation, and the basics of statistics, optimization, and stochastic processes.
 

Web Site Vergil
Department Decision, Risk and Operations
Enrollment 29 students (25 max) as of 5:08PM Saturday, September 7, 2024
Status Full
Subject Decision, Risk & Operations Management
Number B9120
Section 001
Division School of Business
Open To Business, GSAS
Section key 20233DROM9120B001