The paper introduces a new class of parameterized controllers that draw inspiration from Model Predictive Control (MPC). The key idea is to design the controller to resemble a Quadratic Programming (QP) solver of a linear MPC problem, but with the parameters of the controller being trained via Deep Reinforcement Learning (DRL) rather than derived from system models.
The proposed approach addresses the limitations of common DRL controllers with Multi-Layer Perceptron (MLP) or other general neural network architectures, in terms of verifiability and performance guarantees. The learned controllers possess verifiable properties like persistent feasibility and asymptotic stability akin to MPC.
Numerical examples illustrate that the proposed controller empirically matches MPC and MLP controllers in terms of control performance, and has superior robustness against modeling uncertainty and noise. Furthermore, the proposed controller is significantly more computationally efficient compared to MPC and requires fewer parameters to learn than MLP controllers.
Real-world experiments on a vehicle drift maneuvering task demonstrate the potential of these controllers for robotics and other demanding control tasks, despite the high nonlinearity of the system.
The paper also provides theoretical analysis to establish performance guarantees of the learned QP controller, including sufficient conditions for persistent feasibility and asymptotic stability of the closed-loop system.
In un'altra lingua
dal contenuto originale
arxiv.org
Approfondimenti chiave tratti da
by Yiwen Lu,Zis... alle arxiv.org 04-10-2024
https://arxiv.org/pdf/2312.05332.pdfDomande più approfondite