Given the dramatic successes in machine learning over the past half decade, there has been a resurgence of interest in applying learning techniques to continuous control problems in robotics, self-driving cars, and unmanned aerial vehicles. Though such applications appear to be straightforward generalizations of reinforcement learning, it remains unclear which machine learning tools are best equipped to handle decision making, planning, and actuation in highly uncertain dynamic environments.
This tutorial will survey the foundations required to build machine learning systems that reliably act upon the physical world. The primary technical focus will be on numerical optimization tools at the interface of statistical learning and dynamical systems. We will investigate how to learn models of dynamical systems, how to use data to achieve objectives in a timely fashion, how to balance model specification and system controllability, and how to safely acquire new information to improve performance. We will close by listing several exciting open problems that must be solved before we can build robust, reliable learning systems that interact with an uncertain environment.
Speaker: Ben Recht
Time and Location: Tuesday, July 10th, 3:45-6:00pm
Dimitri P. Bertsekas. Dynamic Programming and Optimal Control. 4th edition, volumes 1 (2017) and 2 (2012). Athena Scientific.
Dimitri P. Bertsekas. and John Tsitsiklis. Neuro-dynamic Programming. Athena Scientific, 1996.
Francesco Borrelli, Alberto Bemporad, and Manfred Morari. Predictive Control for Linear and Hybrid Systems. Cambridge, 2017.
Benjamin Recht. “A Tour of Reinforcement Learning: The View from Continuous Control.” arXiv:1806.09460. 2018.
Much of the material in this survey and tutorial was adapted from works on the argmin blog.
Sarah Dean, Horia Mania, Nikolai Matni, Benjamin Recht, and Stephen Tu. “On the Sample Complexity of the Linear Quadratic Regulator.” arXiv:1710.01688. 2017.
Max Simchowitz, Horia Mania, Stephen Tu, Michael I. Jordan, and Benjamin Recht. “Learning Without Mixing: Towards A Sharp Analysis of Linear System Identification.” In COLT 2018.
Stephen Tu and Benjamin Recht. “Least-Squares Temporal Difference Learning for the Linear Quadratic Regulator.” In ICML 2018.
Horia Mania, Aurelia Guy, Benjamin Recht. “Simple Random Search Provides a Competitive Approach to Reinforcement Learning.” arXiv:1803.07055. 2018.
Sarah Dean, Horia Mania, Nikolai Matni, Benjamin Recht, Stephen Tu. Regret Bounds for Robust Adaptive Control of the Linear Quadratic Regulator. arXiv:1805.09388. 2018.
Peter Henderson, Riashat Islam, Philip Bachman, Joelle Pineau, Doina Precup, and David Meger. “Deep Reinforcement Learning that Matters.” arXiv:1709.06560. 2017.
Riashat Islam, Peter Henderson, Maziar Gomrokchi, and Doina Precup. Reproducibility of Benchmarked Deep Reinforcement Learning Tasks for Continuous Control. arxiv:1708.04133. 2017.
Aravind Rajeswaran, Kendall Lowrey, Emanuel Todorovm and Sham Kakade. “Towards Generalization and Simplicity in Continuous Control.” In NIPS 2017.
Yuval Tassa, Tom Erez, and Emanuel Todorov. “Synthesis and Stabilization of Complex Behaviors through Online Trajectory Optimization.” In IROS 2012. video demo
Tom Erez et al. “An Integrated System for Real-time Model Predictive Control of Humanoid Robots.” In IEEE-RAS. 2013. video demo.
Ugo Rosolia and Francesco Borrelli. Learning Model Predictive Control for Iterative Tasks: A Data-Driven Control Framework. IEEE Transactions on Automatic Control. PP(99). 2017. video demo