{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T01:27:42Z","timestamp":1760059662629,"version":"build-2065373602"},"reference-count":35,"publisher":"MDPI AG","issue":"7","license":[{"start":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T00:00:00Z","timestamp":1751414400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Axioms"],"abstract":"<jats:p>This paper studies an averaged Linear Quadratic Regulator (LQR) problem for a parabolic partial differential equation (PDE), where the system dynamics are affected by uncertain parameters. Instead of assuming a deterministic operator, we model the uncertainty using a probability distribution over a set of possible system dynamics. This approach extends classical optimal control theory by incorporating an averaging framework to account for parameter uncertainty. We establish the existence and uniqueness of the optimal control solution and analyze its convergence as the probability distribution governing the system parameters changes. These results provide a rigorous foundation for solving optimal control problems in the presence of parameter uncertainty. Our findings lay the groundwork for further studies on optimal control in dynamic systems with uncertainty.<\/jats:p>","DOI":"10.3390\/axioms14070512","type":"journal-article","created":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T09:08:44Z","timestamp":1751447324000},"page":"512","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Averaging of Linear Quadratic Parabolic Optimal Control Problem"],"prefix":"10.3390","volume":"14","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2629-0750","authenticated-orcid":false,"given":"Olena","family":"Kapustian","sequence":"first","affiliation":[{"name":"Faculty of Computer Science and Cybernetics, Taras Shevchenko National University of Kyiv, Volodymyrska Street 64, 01601 Kyiv, Ukraine"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4194-402X","authenticated-orcid":false,"given":"Oleksandr","family":"Laptiev","sequence":"additional","affiliation":[{"name":"Faculty of Information Technology, Taras Shevchenko National University of Kyiv, Volodymyrska Street 64, 01601 Kyiv, Ukraine"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-2352-9933","authenticated-orcid":false,"given":"Adalbert","family":"Makarovych","sequence":"additional","affiliation":[{"name":"Faculty of Computer Science and Cybernetics, Taras Shevchenko National University of Kyiv, Volodymyrska Street 64, 01601 Kyiv, Ukraine"}]}],"member":"1968","published-online":{"date-parts":[[2025,7,2]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Evans, L.C. (2010). Partial Differential Equations, American Mathematical Society. [2nd ed.].","DOI":"10.1090\/gsm\/019"},{"key":"ref_2","unstructured":"Anderson, B.D.O., and Moore, J.B. (1989). Optimal Control: Linear Quadratic Methods, Prentice Hall."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"297","DOI":"10.1007\/s11072-010-0078-0","article-title":"Approximate bounded synthesis for one weakly nonlinear boundary-value problem","volume":"12","author":"Kapustyan","year":"2009","journal-title":"Nonlinear Oscil."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"379","DOI":"10.1007\/s00498-021-00294-y","article-title":"Convergence results for an averaged LQR problem with applications to reinforcement learning","volume":"33","author":"Pesare","year":"2021","journal-title":"Math. Control Signals Syst."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Sadovnichiy, V., and Zgurovsky, M. (2019). Approximate optimal regulator for distributed control problem with superposition functional and rapidly oscillating coefficients. Modern Mathematics and Mechanics, Springer.","DOI":"10.1007\/978-3-319-96755-4"},{"key":"ref_6","unstructured":"Sutton, R.S., and Barto, A.G. (2018). Reinforcement Learning: An Introduction, MIT Press. [2nd ed.]."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"19","DOI":"10.1109\/37.126844","article-title":"Reinforcement learning is direct adaptive optimal control","volume":"12","author":"Sutton","year":"1992","journal-title":"IEEE Control Syst."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"253","DOI":"10.1146\/annurev-control-053018-023825","article-title":"A tour of reinforcement learning: The view from continuous control","volume":"2","author":"Recht","year":"2019","journal-title":"Annu. Rev. Control Robot Auton. Syst."},{"key":"ref_9","unstructured":"Atkeson, C.G., and Santamaria, J.C. (1997, January 20\u201325). A comparison of direct and model-based reinforcement learning. Proceedings of the International Conference on Robotics and Automation, Albuquerque, NM, USA."},{"key":"ref_10","unstructured":"Deisenroth, M.P. (2010). Efficient Reinforcement Learning Using Gaussian Processes, KIT Scientific Publishing."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"408","DOI":"10.1109\/TPAMI.2013.218","article-title":"Gaussian processes for data-efficient learning in robotics and control","volume":"37","author":"Deisenroth","year":"2013","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_12","unstructured":"Gal, Y., McAllister, R., and Rasmussen, C.E. (2016, January 24). Improving PILCO with Bayesian neural network dynamics models. Proceedings of the ICML Workshop on Data-Efficient Machine Learning, New York, NY, USA."},{"key":"ref_13","first-page":"12519","article-title":"When to trust your model: Model-based policy optimization","volume":"32","author":"Janner","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_14","unstructured":"Kamthe, S., and Deisenroth, M. (2018, January 9\u201311). Data-efficient reinforcement learning with probabilistic model predictive control. Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), Playa Blanca, Lanzarote, Spain."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Chowdhary, G., Kingravi, H.A., How, J.P., and Vela, P.A. (2013, January 10\u201313). A Bayesian nonparametric approach to adaptive control using Gaussian processes. Proceedings of the IEEE Conference Decision Control (CDC), Florence, Italy.","DOI":"10.2514\/6.2013-4692"},{"key":"ref_16","first-page":"4754","article-title":"Deep reinforcement learning in a handful of trials using probabilistic dynamics models","volume":"31","author":"Chua","year":"2018","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_17","unstructured":"Wang, T., Bao, X., Clavera, I., Hoang, J., Wen, Y., Langlois, E., Zhang, S., Zhang, G., Abbeel, P., and Ba, J. (2019). Benchmarking Model-Based Reinforcement Learning. arXiv."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"24","DOI":"10.1016\/j.sysconle.2018.09.011","article-title":"A model for system uncertainty in reinforcement learning","volume":"122","author":"Murray","year":"2018","journal-title":"Syst. Control Lett."},{"key":"ref_19","unstructured":"Murray, R., and Palladino, M. (2019, January 11\u201313). Modelling uncertainty in reinforcement learning. Proceedings of the IEEE Conference on Decision and Control (CDC), Nice, France."},{"key":"ref_20","first-page":"2093","article-title":"Necessary optimality conditions for average cost minimization problems","volume":"24","author":"Bettiol","year":"2019","journal-title":"Discrete Contin. Dyn. Syst. B"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"659","DOI":"10.1007\/s11228-016-0364-9","article-title":"Necessary conditions for adverse control problems expressed by relaxed derivatives","volume":"24","author":"Palladino","year":"2016","journal-title":"Set-Valued Var. Anal."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"1251","DOI":"10.2514\/1.G000505","article-title":"Riemann\u2013Stieltjes optimal control problems for uncertain dynamic systems","volume":"38","author":"Ross","year":"2015","journal-title":"J. Guid. Control Dyn."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"785","DOI":"10.5802\/afst.1511","article-title":"From averaged to simultaneous controllability","volume":"25","author":"Zuazua","year":"2016","journal-title":"Ann. Fac. Sci. Toulouse Math."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"3077","DOI":"10.1016\/j.automatica.2014.10.054","article-title":"Averaged control","volume":"50","author":"Zuazua","year":"2014","journal-title":"Automatica"},{"key":"ref_25","first-page":"70","article-title":"Algorithms for synthesis of functionally stable wireless sensor network","volume":"9","author":"Barabash","year":"2025","journal-title":"Adv. Inf. Syst."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"219","DOI":"10.1162\/089976600300015961","article-title":"Reinforcement learning in continuous time and space","volume":"12","author":"Doya","year":"2000","journal-title":"Neural Comput."},{"key":"ref_27","first-page":"916","article-title":"Integral reinforcement learning for continuous-time input-affine nonlinear systems with simultaneous invariant explorations","volume":"26","author":"Lee","year":"2014","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1109\/MCAS.2009.933854","article-title":"Reinforcement learning and adaptive dynamic programming for feedback control","volume":"9","author":"Lewis","year":"2009","journal-title":"IEEE Circuits Syst. Mag."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"265","DOI":"10.1023\/A:1007686309208","article-title":"A study of reinforcement learning in the continuous case by the means of viscosity solutions","volume":"40","author":"Munos","year":"2000","journal-title":"Mach. Learn."},{"key":"ref_30","unstructured":"Temam, R. (2013). Infinite-Dimensional Dynamical Systems in Mechanics and Physics, Springer Science & Business Media."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Villani, C. (2009). Optimal Transport: Old and New, Springer.","DOI":"10.1007\/978-3-540-71050-9"},{"key":"ref_32","unstructured":"Lions, J.L. (1988). Contr\u00f4labilit\u00e9 Exacte, Perturbations et Stabilisation de Syst\u00e8mes Distribu\u00e9s, Masson."},{"key":"ref_33","unstructured":"Fleming, W.H., and Soner, H.M. (2006). Controlled Markov Processes and Viscosity Solutions, Springer. [2nd ed.]."},{"key":"ref_34","first-page":"81","article-title":"Analysis of socio-economic determinants of youth employment using machine learning methods","volume":"6","author":"Makarovych","year":"2024","journal-title":"Acta Acad. Beregsasiensis Econ."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"66","DOI":"10.24144\/2616-7700.2024.44(1).66-82","article-title":"SingleStoreDB connector for Apache Beam","volume":"44","author":"Glebena","year":"2024","journal-title":"Sci. Bull. Uzhhorod Univ. Ser. Math. Inf."}],"container-title":["Axioms"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2075-1680\/14\/7\/512\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T18:03:12Z","timestamp":1760032992000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2075-1680\/14\/7\/512"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,2]]},"references-count":35,"journal-issue":{"issue":"7","published-online":{"date-parts":[[2025,7]]}},"alternative-id":["axioms14070512"],"URL":"https:\/\/doi.org\/10.3390\/axioms14070512","relation":{},"ISSN":["2075-1680"],"issn-type":[{"type":"electronic","value":"2075-1680"}],"subject":[],"published":{"date-parts":[[2025,7,2]]}}}