Syllabus
Course Syllabus
PDF copy of the syllabus PDF
Module 1: State Estimation (11 lectures)
- Background on probability
- Topics: Markov chains, Hidden Markov Models, Kalman Filter, Extended and Unscented Kalman Filter, particle filters, occupancy grids, transformations
Module 2: Control and Planning (5-6 lectures)
- Background on linear control, dynamic programming
- Topics: Markov Decision Processes, Value and Policy Iteration, Bellman equation, Linear Quadratic Regulator, Linear Quadratic Gaussian, Sampling-based motion planning
Module 3: Reinforcement Learning for Robotics (7 lectures)
- Background on deep learning and optimization
- Topics: Imitation Learning, Policy gradient, Q-Learning, Inverse RL, Model-based RL, Offline RL
Module 4: Miscellaneous topics (2 lecture)
- Meta-Learning, Sim2Real
Tentative schedule
Lecture | Topic | Notes |
---|---|---|
1 | Introduction | HW 0 out (not graded) |
2 | Background on probability | HW 1 out |
3 | Markov Chains | |
4 | Hidden Markov Models I | |
5 | Hidden Markov Models II | |
6 | Kalman Filter | HW 1 due |
7 | Extended Kalman Filter | |
8 | Unscented Kalman Filter | HW 2 out |
9 | Particle Filter | |
10 | Rigid Transforms, Quaternions | |
11 | Occupancy Grids | Summary on Lec 4-10 |
12 | Dynamic Programming, Bellman Equation | HW 2 Due |
13 | Value Iteration | |
14 | Policy iteration | HW 3 out |
15 | Background on Linear Control, LQR | |
16 | LQG, Iterated LQR | |
17 | Midterm | |
18 | Sampling Based Motion Planning | HW 3 Due |
19 | Optimization, Imitation Learning | |
20 | Policy Gradient | |
21 | Tabular Q-Learning | |
22 | Continuous Q-Learning | HW 4 out |
23 | Inverse RL, Model-based RL | |
24 | Offline RL | |
25 | Deep RL | HW 4 due |
26 | Closing topics |