In this paper, we address the problem of on-policy data-driven linear quadratic optimal control for continuous-time single-input single-output systems. Assuming that the plant is minimum phase and has relative degree one, we propose model reference adaptive reinforcement learning – an approach with theoretical guarantees that combines learning and model reference adaptive control. The developed algorithm features an adaptive output-feedback controller that tracks a parameter-varying reference model, whose behavior is shaped by a discrete-time optimizer. For the resulting hybrid closed-loop system, we establish semi-global boundedness of the solutions and show that, under persistency of excitation induced by a dither signal, the applied policy converges to the optimal one.

Bosso, A., Borghesi, M., Serrani, A., Notarstefano, G., Teel, A.R. (2025). On-Policy Data-Driven Linear Quadratic Optimal Control of SISO Systems via Model Reference Adaptive Reinforcement Learning. New York : IEEE [10.1109/cdc57313.2025.11312870].

On-Policy Data-Driven Linear Quadratic Optimal Control of SISO Systems via Model Reference Adaptive Reinforcement Learning

Bosso, Alessandro
Primo
;
Borghesi, Marco
Secondo
;
Serrani, Andrea;Notarstefano, Giuseppe
Penultimo
;
2025

Abstract

In this paper, we address the problem of on-policy data-driven linear quadratic optimal control for continuous-time single-input single-output systems. Assuming that the plant is minimum phase and has relative degree one, we propose model reference adaptive reinforcement learning – an approach with theoretical guarantees that combines learning and model reference adaptive control. The developed algorithm features an adaptive output-feedback controller that tracks a parameter-varying reference model, whose behavior is shaped by a discrete-time optimizer. For the resulting hybrid closed-loop system, we establish semi-global boundedness of the solutions and show that, under persistency of excitation induced by a dither signal, the applied policy converges to the optimal one.
2025
2025 IEEE 64th Conference on Decision and Control (CDC)
2642
2647
Bosso, A., Borghesi, M., Serrani, A., Notarstefano, G., Teel, A.R. (2025). On-Policy Data-Driven Linear Quadratic Optimal Control of SISO Systems via Model Reference Adaptive Reinforcement Learning. New York : IEEE [10.1109/cdc57313.2025.11312870].
Bosso, Alessandro; Borghesi, Marco; Serrani, Andrea; Notarstefano, Giuseppe; Teel, Andrew R.
File in questo prodotto:
File Dimensione Formato  
CDC2025_MR-ARL_SISO.pdf

accesso aperto

Tipo: Postprint / Author's Accepted Manuscript (AAM) - versione accettata per la pubblicazione dopo la peer-review
Licenza: Creative commons
Dimensione 481.52 kB
Formato Adobe PDF
481.52 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1042935
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact