Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/3327144.3327332guideproceedingsArticle/Chapter ViewAbstractPublication PagesnipsConference Proceedingsconference-collections
Article
Free access

Regret bounds for robust adaptive control of the linear quadratic regulator

Published: 03 December 2018 Publication History

Abstract

We consider adaptive control of the Linear Quadratic Regulator (LQR), where an unknown linear system is controlled subject to quadratic costs. Leveraging recent developments in the estimation of linear systems and in robust controller synthesis, we present the first provably polynomial time algorithm that provides high probability guarantees of sub-linear regret on this problem. We further study the interplay between regret minimization and parameter estimation by proving a lower bound on the expected regret in terms of the exploration schedule used by any algorithm. Finally, we conduct a numerical study comparing our robust adaptive algorithm to other methods from the adaptive LQR literature, and demonstrate the flexibility of our proposed method by extending it to a demand forecasting problem subject to state constraints.

References

[1]
Yasin Abbasi-Yadkori and Csaba Szepesvári. Regret Bounds for the Adaptive Control of Linear Quadratic Systems. In Conference on Learning Theory, 2011.
[2]
Yasin Abbasi-Yadkori and Csaba Szepesvári. Bayesian Optimal Control of Smoothly Parameterized Systems: The Lazy Posterior Sampling Algorithm. In Conference on Uncertainty in Artificial Intelligence, 2015.
[3]
Yasin Abbasi-Yadkori, Nevena Lazic, and Csaba Szepesvári. Model-Free Linear Quadratic Control via Reduction to Expert Prediction. arXiv:1804.06021, 2018.
[4]
Marc Abeille and Alessandro Lazaric. Thompson Sampling for Linear-Quadratic Control Problems. In AISTATS, 2017.
[5]
S. Bittanti and M. C. Campi. Adaptive control of linear time invariant systems: the "bet on the best" principle. Communications in Information and Systems, 6(4), 2006.
[6]
Sarah Dean, Horia Mania, Nikolai Matni, Benjamin Recht, and Stephen Tu. On the Sample Complexity of the Linear Quadratic Regulator. arXiv:1710.01688, 2017.
[7]
Sarah Dean, Horia Mania, Nikolai Matni, Benjamin Recht, and Stephen Tu. Regret Bounds for Robust Adaptive Control of the Linear Quadratic Regulator. arXiv:1805.09388, 2018.
[8]
Mohamad Kazem Shirani Faradonbeh, Ambuj Tewari, and George Michailidis. Finite Time Analysis of Optimal Adaptive Policies for Linear-Quadratic Systems. arXiv:1711.07230, 2017.
[9]
Maryam Fazel, Rong Ge, Sham M. Kakade, and Mehran Mesbahi. Global Convergence of Policy Gradient Methods for the Linear Quadratic Regulator. In International Conference on Machine Learning, 2018.
[10]
Claude-Nicolas Fiechter. PAC Adaptive Control of Linear Systems. In Conference on Learning Theory, 1997.
[11]
Morteza Ibrahimi, Adel Javanmard, and Benjamin Van Roy. Efficient Reinforcement Learning for High Dimensional Linear Quadratic Systems. In Neural Information Processing Systems, 2012.
[12]
Petros A Ioannou and Jing Sun. Robust adaptive control, volume 1. PTR Prentice-Hall Upper Saddle River, NJ, 1996.
[13]
Miroslav Krstic, Ioannis Kanellakopoulos, and Peter V Kokotovic. Nonlinear and adaptive control design. Wiley, 1995.
[14]
Nikolai Matni, Yuh-Shyang Wang, and James Anderson. Scalable system level synthesis for virtually localizable systems. In IEEE Conference on Decision and Control, 2017.
[15]
Ian Osband and Benjamin Van Roy. Posterior Sampling for Reinforcement Learning Without Episodes. arXiv:1608.02731, 2016.
[16]
Yi Ouyang, Mukul Gagrani, and Rahul Jain. Learning-based Control of Unknown Linear Systems with Thompson Sampling. arXiv:1709.04047, 2017.
[17]
Max Simchowitz, Horia Mania, Stephen Tu, Michael I. Jordan, and Benjamin Recht. Learning Without Mixing: Towards A Sharp Analysis of Linear System Identification. In Conference on Learning Theory, 2018.
[18]
Stephen Tu and Benjamin Recht. Least-Squares Temporal Difference Learning for the Linear Quadratic Regulator. In International Conference on Machine Learning, 2018.
[19]
Yuh-Shyang Wang, Nikolai Matni, and John C Doyle. A System Level Approach to Controller Synthesis. arXiv:1610.04815, 2016.
[20]
K. Zhou, J. C. Doyle, and K. Glover. Robust and Optimal Control. 1995.

Cited By

View all
  • (2022)Dynamic Regret Minimization for Control of Non-stationary Linear Dynamical SystemsProceedings of the ACM on Measurement and Analysis of Computing Systems10.1145/35080296:1(1-72)Online publication date: 28-Feb-2022

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Guide Proceedings
NIPS'18: Proceedings of the 32nd International Conference on Neural Information Processing Systems
December 2018
11021 pages

Publisher

Curran Associates Inc.

Red Hook, NY, United States

Publication History

Published: 03 December 2018

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)43
  • Downloads (Last 6 weeks)9
Reflects downloads up to 10 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2022)Dynamic Regret Minimization for Control of Non-stationary Linear Dynamical SystemsProceedings of the ACM on Measurement and Analysis of Computing Systems10.1145/35080296:1(1-72)Online publication date: 28-Feb-2022

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media