{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T10:41:05Z","timestamp":1768560065505,"version":"3.49.0"},"reference-count":31,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Automat. Contr."],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1109\/tac.2024.3465566","type":"journal-article","created":{"date-parts":[[2024,9,20]],"date-time":"2024-09-20T17:28:12Z","timestamp":1726853292000},"page":"2106-2121","source":"Crossref","is-referenced-by-count":2,"title":["Policy Algebraic Equation for the Discrete-Time Linear Quadratic Regulator Problem"],"prefix":"10.1109","volume":"70","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4525-4656","authenticated-orcid":false,"given":"Mario","family":"Sassano","sequence":"first","affiliation":[{"name":"Dipartimento di Ingegneria Civile e Ingegneria Informatica, Universit&#x00E0; di Roma &#x201C;Tor Vergata&#x201D;, Roma, Italy"}]}],"member":"263","reference":[{"key":"ref1","volume-title":"Optimal Control: Linear Quadratic Methods","author":"Anderson","year":"1990"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1115\/1.1399382"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1137\/0307007"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-76755-5"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/tsmc.1979.4310157"},{"key":"ref6","volume-title":"Linear Optimal Control Systems","author":"Kwakernaak","year":"1972"},{"issue":"5","key":"ref7","first-page":"430","article-title":"The discrete Riccati equation of optimal control","volume":"8","author":"Kuera","year":"1972","journal-title":"Kybernetika"},{"key":"ref8","volume-title":"Dynamic Programming and Optimal Control","author":"Bertsekas","year":"2005"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.2307\/j.ctvcm4g0s"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1968.1098829"},{"key":"ref11","volume-title":"Reinforcement Learning: An Introduction","author":"Sutton","year":"2018"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1971.1099755"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2022.3199211"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCB.2008.926614"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/MCAS.2009.933854"},{"key":"ref16","doi-asserted-by":"crossref","DOI":"10.1016\/j.automatica.2021.109687","article-title":"A novel adaptive dynamic programming based on tracking error for nonlinear discrete-time systems","volume":"129","author":"Li","year":"2021","journal-title":"Automatica"},{"issue":"5","key":"ref17","doi-asserted-by":"crossref","first-page":"878","DOI":"10.1016\/j.automatica.2010.02.018","article-title":"Online actorcritic algorithm to solve the continuous-time infinite horizon optimal control problem","volume":"46","author":"Vamvoudakis","year":"2010","journal-title":"Automatica"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.2022.3145693"},{"key":"ref19","volume-title":"Neuro-Dynamic Programming","author":"Bertsekas","year":"1996"},{"key":"ref20","article-title":"LQR through the lens of first order methods: Discrete-time case","author":"Bu","year":"2019"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/LCSYS.2023.3271594"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/LCSYS.2024.3382439"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2020.3021385"},{"key":"ref24","volume-title":"The Theory of Matrices","author":"Gantmacher","year":"1959"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1977.1101435"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1983.1103160"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1137\/0717046"},{"key":"ref28","article-title":"Controlgym: Large-scale safety-critical control environments for benchmarking reinforcement learning algorithms","author":"Zhang","year":"2023"},{"key":"ref29","volume-title":"Robust and Optimal Control","author":"Zhou","year":"1996"},{"key":"ref30","doi-asserted-by":"crossref","DOI":"10.1007\/978-1-4471-0507-7","volume-title":"L2-Gain and Passivity Techniques in Nonlinear Control","author":"van der Schaft","year":"2000"},{"issue":"6","key":"ref31","doi-asserted-by":"crossref","first-page":"770","DOI":"10.1109\/9.256331","article-title":"$L_{2}$-gain analysis of nonlinear systems and nonlinear state-feedback $H_{\\infty }$ control","volume":"37","author":"Van der Schaft","year":"1992","journal-title":"IEEE Trans. Autom. Control"}],"container-title":["IEEE Transactions on Automatic Control"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/9\/10945509\/10685125.pdf?arnumber=10685125","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,29]],"date-time":"2025-03-29T09:22:08Z","timestamp":1743240128000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10685125\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4]]},"references-count":31,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/tac.2024.3465566","relation":{},"ISSN":["0018-9286","1558-2523","2334-3303"],"issn-type":[{"value":"0018-9286","type":"print"},{"value":"1558-2523","type":"electronic"},{"value":"2334-3303","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4]]}}}