Implicit regularization and momentum algorithms in nonlinearly parameterized adaptive control and prediction

Stable concurrent learning and control of dynamical systems is the subject of adaptive control. Despite being an established field with many practical applications and a rich theory, much of the development in adaptive control for nonlinear systems revolves around a few key algorithms. By exploiting...

Full description

Bibliographic Details
Main Authors: Boffi, N.M (Author), Slotine, J.-J.E (Author)
Format: Article
Language:English
Published: MIT Press Journals 2021
Subjects:
Online Access:View Fulltext in Publisher
LEADER 03100nam a2200493Ia 4500
001 10.1162-neco_a_01360
008 220427s2021 CNT 000 0 und d
020 |a 08997667 (ISSN) 
245 1 0 |a Implicit regularization and momentum algorithms in nonlinearly parameterized adaptive control and prediction 
260 0 |b MIT Press Journals  |c 2021 
856 |z View Fulltext in Publisher  |u https://doi.org/10.1162/neco_a_01360 
520 3 |a Stable concurrent learning and control of dynamical systems is the subject of adaptive control. Despite being an established field with many practical applications and a rich theory, much of the development in adaptive control for nonlinear systems revolves around a few key algorithms. By exploiting strong connections between classical adaptive nonlinear control techniques and recent progress in optimization and machine learning, we show that there exists considerable untapped potential in algorithm development for both adaptive nonlinear control and adaptive dynamics prediction. We begin by introducing first-order adaptation laws inspired by natural gradient descent and mirror descent. We prove that when there are multiple dynamics consistent with the data, these non-Euclidean adaptation laws implicitly regularize the learned model. Local geometry imposed during learning thus may be used to select parameter vectors—out of the many that will achieve perfect tracking or prediction—for desired properties such as sparsity. We apply this result to regularized dynamics predictor and observer design, and as concrete examples, we consider Hamiltonian systems, Lagrangian systems, and recurrent neural networks. We subsequently develop a variational formalism based on the Bregman Lagrangian. We show that its Euler Lagrange equations lead to natural gradient and mirror descent-like adaptation laws with momentum, and we recover their first-order analogues in the infinite friction limit. We illustrate our analyses with simulations demonstrating our theoretical results. © 2021 Massachusetts Institute of Technology. 
650 0 4 |a Adaptive Control 
650 0 4 |a Adaptive control systems 
650 0 4 |a Adaptive dynamics 
650 0 4 |a Adaptive nonlinear control 
650 0 4 |a algorithm 
650 0 4 |a Algorithm development 
650 0 4 |a article 
650 0 4 |a Concurrency control 
650 0 4 |a Control theory 
650 0 4 |a Dynamical systems 
650 0 4 |a Dynamics 
650 0 4 |a Equations of motion 
650 0 4 |a Euler-Lagrange equations 
650 0 4 |a Forecasting 
650 0 4 |a geometry 
650 0 4 |a Gradient methods 
650 0 4 |a Hamiltonian systems 
650 0 4 |a Hamiltonians 
650 0 4 |a human 
650 0 4 |a Lagrange multipliers 
650 0 4 |a Lagrangian system 
650 0 4 |a Learning systems 
650 0 4 |a machine learning 
650 0 4 |a Mirrors 
650 0 4 |a Parameter vectors 
650 0 4 |a prediction 
650 0 4 |a recurrent neural network 
650 0 4 |a Recurrent neural networks 
650 0 4 |a simulation 
700 1 |a Boffi, N.M.  |e author 
700 1 |a Slotine, J.-J.E.  |e author 
773 |t Neural Computation