Revisiting LQR Control From the Perspective of Receding-Horizon Policy Gradient

Xiangyuan Zhang, Tamer Basar

Research output: Contribution to journalArticlepeer-review

Abstract

We revisit in this letter the discrete-time linear quadratic regulator (LQR) problem from the perspective of receding-horizon policy gradient (RHPG), a newly developed model-free learning framework for control applications. We provide a fine-grained sample complexity analysis for RHPG to learn a control policy that is both stabilizing and ϵ-close to the optimal LQR solution, and our algorithm does not require knowing a stabilizing control policy for initialization. Combined with the recent application of RHPG in learning the Kalman filter, we demonstrate the general applicability of RHPG in linear control and estimation with streamlined analyses.

Original languageEnglish (US)
Pages (from-to)1664-1669
Number of pages6
JournalIEEE Control Systems Letters
Volume7
DOIs
StatePublished - 2023

Keywords

  • Optimal control
  • optimization
  • reinforcement learning
  • sample complexity

ASJC Scopus subject areas

  • Control and Optimization
  • Control and Systems Engineering

Fingerprint

Dive into the research topics of 'Revisiting LQR Control From the Perspective of Receding-Horizon Policy Gradient'. Together they form a unique fingerprint.

Cite this