TY - GEN
T1 - On Imitation Learning of Linear Control Policies
T2 - 2021 American Control Conference, ACC 2021
AU - Havens, Aaron
AU - Hu, Bin
N1 - Publisher Copyright:
© 2021 American Automatic Control Council.
PY - 2021/5/25
Y1 - 2021/5/25
N2 - When applying imitation learning techniques to fit a policy from expert demonstrations, one can take advantage of prior stability/robustness assumptions on the expert's policy and incorporate such control-theoretic prior knowledge explicitly into the learning process. In this paper, we formulate the imitation learning of linear policies as a constrained optimization problem, and present efficient methods which can be used to enforce stability and robustness constraints during the learning processes. Specifically, we show that one can guarantee the closed-loop stability and robustness by posing linear matrix inequality (LMI) constraints on the fitted policy. Then both the projected gradient descent method and the alternating direction method of multipliers (ADMM) method can be applied to solve the resultant constrained policy fitting problem. Finally, we provide numerical results to demonstrate the effectiveness of our methods in producing linear polices with various stability and robustness guarantees.
AB - When applying imitation learning techniques to fit a policy from expert demonstrations, one can take advantage of prior stability/robustness assumptions on the expert's policy and incorporate such control-theoretic prior knowledge explicitly into the learning process. In this paper, we formulate the imitation learning of linear policies as a constrained optimization problem, and present efficient methods which can be used to enforce stability and robustness constraints during the learning processes. Specifically, we show that one can guarantee the closed-loop stability and robustness by posing linear matrix inequality (LMI) constraints on the fitted policy. Then both the projected gradient descent method and the alternating direction method of multipliers (ADMM) method can be applied to solve the resultant constrained policy fitting problem. Finally, we provide numerical results to demonstrate the effectiveness of our methods in producing linear polices with various stability and robustness guarantees.
UR - http://www.scopus.com/inward/record.url?scp=85111906365&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85111906365&partnerID=8YFLogxK
U2 - 10.23919/ACC50511.2021.9483019
DO - 10.23919/ACC50511.2021.9483019
M3 - Conference contribution
AN - SCOPUS:85111906365
T3 - Proceedings of the American Control Conference
SP - 882
EP - 887
BT - 2021 American Control Conference, ACC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 25 May 2021 through 28 May 2021
ER -