{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,24]],"date-time":"2025-08-24T00:01:09Z","timestamp":1755993669050,"version":"3.44.0"},"reference-count":157,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,7,8]],"date-time":"2025-07-08T00:00:00Z","timestamp":1751932800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,7,8]],"date-time":"2025-07-08T00:00:00Z","timestamp":1751932800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,7,8]]},"DOI":"10.23919\/acc63710.2025.11107771","type":"proceedings-article","created":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T18:17:51Z","timestamp":1755800271000},"page":"4776-4784","source":"Crossref","is-referenced-by-count":0,"title":["Structure-Exploiting Reinforcement Learning for Networked Systems"],"prefix":"10.23919","author":[{"given":"Guannan","family":"Qu","sequence":"first","affiliation":[{"name":"Carnegie Mellon University,Department of Electrical and Computer Engineering,Pittsburgh,PA,USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.13140\/RG.2.2.18893.74727"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1038\/nature24270"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1126\/science.aar6404"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-020-03051-4"},{"key":"ref6","first-page":"1329","article-title":"Benchmarking deep reinforcement learning for continuous control","volume-title":"International Conference on Machine Learning","author":"Duan"},{"issue":"1","key":"ref7","first-page":"1334","article-title":"End-to-end training of deep visuomotor policies","volume":"17","author":"Levine","year":"2016","journal-title":"The Journal of Machine Learning Research"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989385"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2018.xiv.049"},{"article-title":"Visual foresight: Model-based deep reinforcement learning for vision-based robotic control","year":"2018","author":"Ebert","key":"ref10"},{"year":"2023","key":"ref11","article-title":"GPT-4 technical report"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.17775\/CSEEJPES.2019.00920"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2022.3154718"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1049\/iet-its.2009.0070"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.3390\/app10114011"},{"issue":"54","key":"ref16","first-page":"1","article-title":"Deep reinforcement learning for swarm systems","volume":"20","author":"H\u00fcttenrauch","year":"2019","journal-title":"Journal of Machine Learning Research"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-92790-5_31"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2916583"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1016\/B978-1-55860-335-6.50027-1"},{"key":"ref20","first-page":"746","article-title":"The dynamics of reinforcement learning in cooperative multiagent systems","volume":"1998","author":"Claus","year":"1998","journal-title":"AAAI\/IAAI"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1016\/S1389-0417(01)00015-8"},{"issue":"Nov","key":"ref22","first-page":"1039","article-title":"Nash Q-learning for general-sum stochastic games","volume":"4","author":"Hu","year":"2003","journal-title":"Journal of machine learning research"},{"key":"ref23","first-page":"242","article-title":"Multiagent reinforcement learning: theoretical framework and an algorithm","volume":"98","author":"Hu","year":"1998","journal-title":"ICML"},{"key":"ref24","first-page":"535","article-title":"An algorithm for distributed reinforcement learning in cooperative multi-agent systems","volume-title":"Proceedings of the Seventeenth International Conference on Machine Learning, ICML \u201900","author":"Lauer"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1016\/S1389-0417(01)00015-8"},{"key":"ref26","article-title":"Multi-agent reinforcement learning: a critical survey","volume-title":"tech. rep.","author":"Shoham","year":"2003"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-005-2631-2"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCC.2007.913919"},{"article-title":"Asynchronous gradient play in zero-sum multi-agent games","volume-title":"The Eleventh International Conference on Learning Representations","author":"Ao","key":"ref29"},{"key":"ref30","first-page":"15353","article-title":"Minimax-optimal multi-agent RL in Markov games with a generative model","volume":"35","author":"Li","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref31","first-page":"33262","article-title":"Learning two-player Markov games: Neural function approximation and correlated equilibrium","volume":"35","author":"Li","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1145\/3490486.3538289"},{"article-title":"A finite-sample analysis of payoff-based independent learning in zero-sum stochastic games","year":"2023","author":"Chen","key":"ref33"},{"key":"ref34","first-page":"5872","article-title":"Fully decentralized multi-agent reinforcement learning with networked agents","volume-title":"International Conference on Machine Learning","author":"Zhang"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2013.2241057"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2014.2368731"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2016.2585302"},{"key":"ref38","first-page":"9672","article-title":"Multi-agent reinforcement learning via double averaging primal-dual optimization","volume-title":"Proceedings of the 32nd International Conference on Neural Information Processing Systems","author":"Wai"},{"article-title":"V-learning\u2013a simple, efficient, decentralized algorithm for multiagent RL","year":"2021","author":"Jin","key":"ref39"},{"key":"ref40","first-page":"5166","article-title":"Independent policy gradient for large-scale Markov potential games: Sharper rates, function approximation, and game-agnostic convergence","volume-title":"International Conference on Machine Learning","author":"Ding"},{"key":"ref41","first-page":"15007","article-title":"On improving model-free algorithms for decentralized multi-agent reinforcement learning","volume-title":"International Conference on Machine Learning","author":"Mao"},{"key":"ref42","first-page":"25779","article-title":"When are offline two-player zero-sum Markov games solvable?","volume":"35","author":"Cui","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref43","first-page":"1337","article-title":"Near-optimal learning of extensive-form games with imperfect information","volume-title":"International Conference on Machine Learning","author":"Bai"},{"article-title":"Model-based reinforcement learning is minimax-optimal for offline zero-sum Markov games","year":"2022","author":"Yan","key":"ref44"},{"key":"ref45","first-page":"9343","article-title":"Regret minimization and convergence to equilibria in general-sum Markov games","volume-title":"International Conference on Machine Learning","author":"Erez"},{"key":"ref46","first-page":"20469","article-title":"I2q: A fully decentralized Q-learning algorithm","volume":"35","author":"Jiang","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref47","first-page":"25560","article-title":"Finding correlated equilibrium of constrained Markov game: A primal-dual approach","volume":"35","author":"Chen","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref48","first-page":"315","article-title":"Provably efficient generalized Lagrangian policy optimization for safe multi-agent reinforcement learning","volume-title":"Learning for Dynamics and Control Conference","author":"Ding"},{"key":"ref49","first-page":"573","article-title":"Decentralized cooperative reinforcement learning with hierarchical information structure","volume-title":"International Conference on Algorithmic Learning Theory","author":"Kao"},{"article-title":"Finite-sample analysis of decentralized Q-learning for stochastic games","year":"2021","author":"Gao","key":"ref50"},{"article-title":"A survey on large-population systems and scalable multi-agent reinforcement learning","year":"2022","author":"Cui","key":"ref51"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/CDC51059.2022.9993175"},{"article-title":"Partially observable multi-agent RL with (quasi-) efficiency: The blessing of information sharing","volume-title":"ICML","author":"Liu","key":"ref53"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-60990-0_12"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1137\/20M1311971"},{"article-title":"Primal-dual distributed temporal difference learning","year":"2020","author":"Lee","key":"ref56"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/TCNS.2021.3078100"},{"key":"ref58","article-title":"Communication-efficient distributed learning via lazily aggregated quantized gradients","author":"Sun","year":"2019","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2016.2585302"},{"key":"ref60","first-page":"2159","article-title":"Near-optimal reinforcement learning with self-play","volume":"33","author":"Bai","year":"2020","journal-title":"Advances in neural information processing systems"},{"article-title":"When can we learn general-sum Markov games with a large number of players sample-efficiently?","volume-title":"International Conference on Learning Representations","author":"Song","key":"ref61"},{"article-title":"Breaking the curse of multiagency: Provably efficient decentralized multi-agent RL with function approximation","year":"2023","author":"Wang","key":"ref62"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1007\/s13235-021-00420-0"},{"key":"ref64","first-page":"11739","article-title":"Provably efficient offline multi-agent reinforcement learning via strategy-wise bonus","volume":"35","author":"Cui","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref65","first-page":"21886","article-title":"Policy optimization for Markov games: Unified framework and faster convergence","volume":"35","author":"Zhang","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref66","first-page":"4180","article-title":"The complexity of Markov equilibrium in stochastic games","volume-title":"The Thirty Sixth Annual Conference on Learning Theory","author":"Daskalakis"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1287\/moor.2022.1268"},{"key":"ref68","first-page":"10251","article-title":"The power of exploiter: Provable multi-agent RL in large state spaces","volume-title":"International Conference on Machine Learning","author":"Jin"},{"article-title":"Towards general function approximation in zero-sum Markov games","volume-title":"International Conference on Learning Representations","author":"Huang","key":"ref69"},{"article-title":"Decentralized optimistic hyperpolicy mirror descent: Provably no-regret learning in Markov games","volume-title":"The Eleventh International Conference on Learning Representations","author":"Zhan","key":"ref70"},{"key":"ref71","first-page":"24496","article-title":"A self-play posterior sampling algorithm for zero-sum Markov games","volume-title":"International Conference on Machine Learning","author":"Xiong"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1214\/24-aos2483"},{"article-title":"Representation learning for low-rank general-sum Markov games","volume-title":"The Eleventh International Conference on Learning Representations","author":"Ni","key":"ref73"},{"key":"ref74","first-page":"6379","article-title":"Multi-agent actor-critic for mixed cooperative-competitive environments","author":"Lowe","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11794"},{"key":"ref76","article-title":"Maven: Multi-agent variational exploration","volume":"32","author":"Mahajan","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref77","first-page":"2085","article-title":"Value-decomposition networks for cooperative multi-agent learning based on team reward","volume-title":"Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems","author":"Sunehag"},{"article-title":"Learning to factorize with transformation for cooperative multi-agent reinforcement learning [c]","volume-title":"Proceedings of the 31st International Conference on Machine Learning, Proceedings of Machine Learning Research","author":"WJKDE","key":"ref78"},{"issue":"1","key":"ref79","first-page":"7234","article-title":"Monotonic value function factorisation for deep multi-agent reinforcement learning","volume":"21","author":"Rashid","year":"2020","journal-title":"The Journal of Machine Learning Research"},{"article-title":"RODE: Learning roles to decompose multi-agent tasks","volume-title":"International Conference on Learning Representations","author":"Wang","key":"ref80"},{"article-title":"QPLEX: Duplex dueling multi-agent Q-learning","volume-title":"International Conference on Learning Representations","author":"Wang","key":"ref81"},{"key":"ref82","doi-asserted-by":"publisher","DOI":"10.1287\/stsy.2021.0081"},{"key":"ref83","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2014.2298140"},{"key":"ref84","doi-asserted-by":"publisher","DOI":"10.1109\/SmartGridComm.2013.6687943"},{"key":"ref85","doi-asserted-by":"publisher","DOI":"10.1109\/ALLERTON.2014.7028508"},{"key":"ref86","doi-asserted-by":"publisher","DOI":"10.1109\/TPWRS.2022.3176525"},{"key":"ref87","first-page":"581","article-title":"Stable reinforcement learning with unbounded state space","volume-title":"Learning for Dynamics and Control","author":"Shah","year":"2020"},{"key":"ref88","article-title":"A Lyapunov-based approach to safe reinforcement learning","author":"Chow","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2020.xvi.088"},{"key":"ref90","article-title":"Safe model-based reinforcement learning with stability guarantees","volume":"30","author":"Berkenkamp","year":"2017","journal-title":"Advances in neural information processing systems"},{"article-title":"Temporal logic guided safe reinforcement learning using control barrier functions","year":"2019","author":"Li","key":"ref91"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33013387"},{"key":"ref93","doi-asserted-by":"publisher","DOI":"10.1201\/9781315140223"},{"key":"ref94","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2002.1184811"},{"key":"ref95","doi-asserted-by":"publisher","DOI":"10.1016\/j.sysconle.2004.08.007"},{"key":"ref96","first-page":"22","article-title":"Constrained policy optimization","volume-title":"International conference on machine learning","author":"Achiam"},{"article-title":"Reward constrained policy optimization","volume-title":"International Conference on Learning Representations","author":"Tessler","key":"ref97"},{"key":"ref98","article-title":"Constrained reinforcement learning has zero duality gap","volume":"32","author":"Paternain","year":"2019","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref99","first-page":"8378","article-title":"Natural policy gradient primal-dual method for constrained Markov decision processes","volume":"33","author":"Ding","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref100","first-page":"3304","article-title":"Provably efficient safe exploration via primal-dual policy optimization","volume-title":"International Conference on Artificial Intelligence and Statistics","author":"Ding"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2021\/614"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2022.3152724"},{"key":"ref103","doi-asserted-by":"publisher","DOI":"10.23919\/acc53348.2022.9867805"},{"key":"ref104","first-page":"3274","article-title":"Triple-q: A model-free algorithm for constrained reinforcement learning with sublinear regret and zero constraint violation","volume-title":"International Conference on Artificial Intelligence and Statistics","author":"Wei"},{"key":"ref105","doi-asserted-by":"publisher","DOI":"10.1109\/CISS56502.2023.10089763"},{"key":"ref106","doi-asserted-by":"publisher","DOI":"10.1109\/LCSYS.2022.3231970"},{"article-title":"State-wise constrained policy optimization","year":"2023","author":"Zhao","key":"ref107"},{"key":"ref108","doi-asserted-by":"publisher","DOI":"10.23919\/ECC.2019.8796030"},{"key":"ref109","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2016.2638961"},{"key":"ref110","doi-asserted-by":"publisher","DOI":"10.1109\/CDC45484.2021.9682859"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.1177\/0278364920949931"},{"key":"ref112","doi-asserted-by":"publisher","DOI":"10.1109\/tro.2022.3232542"},{"key":"ref113","first-page":"466","article-title":"The Lyapunov neural network: Adaptive stability certification for safe learning of dynamical systems","volume-title":"Conference on Robot Learning","author":"Richards"},{"key":"ref114","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-72016-2_20"},{"key":"ref115","article-title":"Learning stable deep dynamics models","volume":"32","author":"Kolter","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref116","doi-asserted-by":"publisher","DOI":"10.1109\/LCSYS.2020.3005328"},{"key":"ref117","doi-asserted-by":"publisher","DOI":"10.1145\/3447928.3456646"},{"key":"ref118","doi-asserted-by":"publisher","DOI":"10.1145\/3365365.3382222"},{"key":"ref119","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341190"},{"article-title":"Learning safe multi-agent control with decentralized neural barrier certificates","volume-title":"International Conference on Learning Representations","author":"Qin","key":"ref120"},{"key":"ref121","first-page":"654","article-title":"Guaranteeing safety of learned perception modules via measurement-robust control barrier functions","volume-title":"Conference on Robot Learning","author":"Dean"},{"key":"ref122","doi-asserted-by":"publisher","DOI":"10.1109\/CDC51059.2022.9992902"},{"key":"ref123","first-page":"708","article-title":"Learning for safety-critical control with control barrier functions","volume-title":"Learning for Dynamics and Control","author":"Taylor","year":"2020"},{"key":"ref124","doi-asserted-by":"publisher","DOI":"10.23919\/ACC50511.2021.9483420"},{"key":"ref125","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2020.XVI.098"},{"key":"ref126","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3156648"},{"key":"ref127","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2021.3098436"},{"key":"ref128","doi-asserted-by":"publisher","DOI":"10.1109\/71.963420"},{"key":"ref129","doi-asserted-by":"publisher","DOI":"10.1109\/CDC40024.2019.9029635"},{"key":"ref130","doi-asserted-by":"publisher","DOI":"10.1287\/opre.2021.2226"},{"key":"ref131","doi-asserted-by":"publisher","DOI":"10.1016\/j.arcontrol.2017.09.002"},{"key":"ref132","doi-asserted-by":"publisher","DOI":"10.1145\/1284680.1284681"},{"key":"ref133","doi-asserted-by":"publisher","DOI":"10.1103\/PhysRevE.68.066101"},{"key":"ref134","doi-asserted-by":"publisher","DOI":"10.1007\/s11134-019-09598-y"},{"key":"ref135","doi-asserted-by":"publisher","DOI":"10.1145\/774763.774784"},{"key":"ref136","doi-asserted-by":"publisher","DOI":"10.1287\/moor.24.2.293"},{"key":"ref137","doi-asserted-by":"publisher","DOI":"10.1177\/0278364915581863"},{"key":"ref138","doi-asserted-by":"publisher","DOI":"10.1109\/TCST.2015.2446955"},{"key":"ref139","doi-asserted-by":"publisher","DOI":"10.23919\/acc.2017.7963043"},{"key":"ref140","article-title":"Scalable multi-agent reinforcement learning for networked systems with average reward","volume":"33","author":"Qu","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref141","doi-asserted-by":"publisher","DOI":"10.1145\/3579443"},{"article-title":"Compositional neural certificates for networked dynamical systems","year":"2023","author":"Zhang","key":"ref142"},{"key":"ref143","doi-asserted-by":"publisher","DOI":"10.1016\/0005-1098(96)00051-9"},{"key":"ref144","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2011.06.018"},{"article-title":"Locally interdependent multi-agent mdp: Theoretical framework for decentralized agents with dynamic dependencies","volume-title":"Forty-first International Conference on Machine Learning","author":"DeWeese","key":"ref145"},{"key":"ref146","doi-asserted-by":"publisher","DOI":"10.1109\/ICUAS.2019.8798116"},{"key":"ref147","first-page":"5571","article-title":"Mean field multi-agent reinforcement learning","volume-title":"International Conference on Machine Learning","author":"Yang"},{"key":"ref148","doi-asserted-by":"publisher","DOI":"10.1137\/20M1360700"},{"key":"ref149","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3900139"},{"key":"ref150","doi-asserted-by":"publisher","DOI":"10.1214\/23-aap1949"},{"key":"ref151","doi-asserted-by":"publisher","DOI":"10.1007\/s00498-021-00310-1"},{"issue":"1","key":"ref152","first-page":"5614","article-title":"On the approximation of cooperative heterogeneous multi-agent reinforcement learning (MARL) using mean field control (MFC)","volume":"23","author":"Mondal","year":"2022","journal-title":"The Journal of Machine Learning Research"},{"key":"ref153","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2008.2009515"},{"key":"ref154","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2013.50"},{"article-title":"Representation learning on graphs: Methods and applications","year":"2017","author":"Hamilton","key":"ref155"},{"key":"ref156","doi-asserted-by":"publisher","DOI":"10.1109\/TBDATA.2018.2850013"},{"article-title":"Few-shot learning via learning the representation, provably","year":"2020","author":"Du","key":"ref157"}],"event":{"name":"2025 American Control Conference (ACC)","start":{"date-parts":[[2025,7,8]]},"location":"Denver, CO, USA","end":{"date-parts":[[2025,7,10]]}},"container-title":["2025 American Control Conference (ACC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11107441\/11107442\/11107771.pdf?arnumber=11107771","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T23:57:44Z","timestamp":1755907064000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11107771\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,8]]},"references-count":157,"URL":"https:\/\/doi.org\/10.23919\/acc63710.2025.11107771","relation":{},"subject":[],"published":{"date-parts":[[2025,7,8]]}}}