Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1609/aaai.v38i8.28630guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
research-article

Neural time-reversed generalized riccati equation

Published: 20 February 2024 Publication History

Abstract

Optimal control deals with optimization problems in which variables steer a dynamical system, and its outcome contributes to the objective function. Two classical approaches to solving these problems are Dynamic Programming and the Pontryagin Maximum Principle. In both approaches, Hamiltonian equations offer an interpretation of optimality through auxiliary variables known as costates. However, Hamiltonian equations are rarely used due to their reliance on forward-backward algorithms across the entire temporal domain. This paper introduces a novel neural-based approach to optimal control, with the aim of working forward-in-time. Neural networks are employed not only for implementing state dynamics but also for estimating costate variables. The parameters of the latter network are determined at each time step using a newly introduced local policy referred to as the time-reversed generalized Riccati equation. This policy is inspired by a result discussed in the Linear Quadratic (LQ) problem, which we conjecture stabilizes state dynamics. We support this conjecture by discussing experimental results from a range of optimal control case studies.

References

[1]
Ambrosio, L.; Fusco, N.; and Pallara, D. 2000. Functions of Bounded Variation and Free continuity Problems. Oxford Mathematical Monographs, The Clarendon Press Oxford University Press.
[2]
Bardi, M.; Dolcetta, I. C.; et al. 1997. Optimal control and viscosity solutions of Hamilton-Jacobi-Bellman equations, volume 12. Springer.
[3]
Bertsekas, D. 2019. Reinforcement learning and optimal control. Athena Scientific.
[4]
Betti, A.; Casoni, M.; Gori, M.; Marullo, S.; Melacci, S.; and Tiezzi, M. 2023. Neural Time-Reversed Generalized Riccati Equation. arXiv:2312.09310.
[5]
Betti, A.; Faggi, L.; Gori, M.; Tiezzi, M.; Marullo, S.; Meloni, E.; and Melacci, S. 2022. Continual Learning through Hamilton Equations. In Conference on Lifelong Learning Agents, 201-212. PMLR.
[6]
Courant, R.; and Hilbert, D. 2008. Methods of mathematical physics: partial differential equations. John Wiley & Sons.
[7]
De Lange, M.; Aljundi, R.; Masana, M.; Parisot, S.; Jia, X.; Leonardis, A.; Slabaugh, G.; and Tuytelaars, T. 2021. A continual learning survey: Defying forgetting in classification tasks. IEEE transactions on pattern analysis and machine intelligence, 44(7): 3366-3385.
[8]
Diehl, M.; Bock, H. G.; and Schloder, J. P. 2005. A Real-Time Iteration Scheme for Nonlinear Optimization in Optimal Feedback Control. SIAM Journal on Control and Optimization, 43(5): 1714-1736.
[9]
Evans, L. C. 2022. Partial differential equations, volume 19. American Mathematical Society.
[10]
Gamkrelidze, R.; Pontrjagin, L. S.; and Boltjanskij, V. G. 1964. The mathematical theory of optimal processes. Macmillan Company.
[11]
Garcia, C. E.; Prett, D.M.; and Morari, M. 1989. Model predictive control: Theory and practice—A survey. Automatica, 25(3): 335-348.
[12]
Giaquinta, M.; and Hildebrandt, S. 2013. Calculus of variations II, volume 311. Springer Science & Business Media.
[13]
Hinton, G. 2022. The forward-forward algorithm: Some preliminary investigations. arXiv preprint arXiv:2212.13345.
[14]
Jin, W.; Wang, Z.; Yang, Z.; and Mou, S. 2019. Pontryagin Differentiable Programming: An End-to-End Learning and Control Framework. ArXiv, abs/1912.12970.
[15]
Lewis, F. L.; Vrabie, D.; and Syrmos, V. L. 2012. Optimal control. John Wiley & Sons.
[16]
Mai, Z.; Li, R.; Jeong, J.; Quispe, D.; Kim, H.; and Sanner, S. 2022. Online continual learning in image classification: An empirical survey. Neurocomputing, 469: 28-51.
[17]
Osborne, M. R. 1969. On shooting methods for boundary value problems. Journal of mathematical analysis and applications, 27(2): 417-433.
[18]
Sutton, R. S.; and Barto, A. G. 2018. Reinforcement learning: An introduction. MIT press.
[19]
Zhang, H.; Wang, Z.; and Liu, D. 2014. A Comprehensive Review of Stability Analysis of Continuous-Time Recurrent Neural Networks. IEEE Transactions on Neural Networks and Learning Systems, 25(7): 1229-1262.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Guide Proceedings
AAAI'24/IAAI'24/EAAI'24: Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence
February 2024
23861 pages
ISBN:978-1-57735-887-9

Sponsors

  • Association for the Advancement of Artificial Intelligence

Publisher

AAAI Press

Publication History

Published: 20 February 2024

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 07 Mar 2025

Other Metrics

Citations

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media