{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T12:06:32Z","timestamp":1775736392247,"version":"3.50.1"},"reference-count":37,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"DOI":"10.13039\/100000006","name":"Office of Naval Research","doi-asserted-by":"publisher","award":["HPCM034125HQU"],"award-info":[{"award-number":["HPCM034125HQU"]}],"id":[{"id":"10.13039\/100000006","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000006","name":"Office of Naval Research","doi-asserted-by":"publisher","award":["N00014-18-1-1238"],"award-info":[{"award-number":["N00014-18-1-1238"]}],"id":[{"id":"10.13039\/100000006","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000006","name":"Office of Naval Research","doi-asserted-by":"publisher","award":["N00173-16-1-G905"],"award-info":[{"award-number":["N00173-16-1-G905"]}],"id":[{"id":"10.13039\/100000006","id-type":"DOI","asserted-by":"publisher"}]},{"name":"NASA Space Technology Institute","award":["80NSSC19K1076"],"award-info":[{"award-number":["80NSSC19K1076"]}]},{"name":"POR Campania Italy","award":["A FESR 2014\/2020, ITS for Logistics, awarded to CN"],"award-info":[{"award-number":["A FESR 2014\/2020, ITS for Logistics, awarded to CN"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/access.2022.3148127","type":"journal-article","created":{"date-parts":[[2022,1,31]],"date-time":"2022-01-31T22:15:11Z","timestamp":1643667311000},"page":"15193-15216","source":"Crossref","is-referenced-by-count":17,"title":["A Unifying View of Estimation and Control Using Belief Propagation With Application to Path Planning"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3777-3501","authenticated-orcid":false,"given":"Francesco A. N.","family":"Palmieri","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0565-181X","authenticated-orcid":false,"given":"Krishna R.","family":"Pattipati","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9757-1712","authenticated-orcid":false,"given":"Giovanni Di","family":"Gennaro","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4889-8505","authenticated-orcid":false,"given":"Giovanni","family":"Fioretti","sequence":"additional","affiliation":[]},{"given":"Francesco","family":"Verolla","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3494-2648","authenticated-orcid":false,"given":"Amedeo","family":"Buonanno","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/1143844.1143963"},{"issue":"3","key":"ref2","first-page":"23","article-title":"Probabilistic inference as a model of planned behavior","volume":"23","author":"Toussaint","year":"2009","journal-title":"K\u00fcnstliche Intell."},{"key":"ref3","article-title":"Reinforcement learning and control as probabilistic inference: Tutorial and review","author":"Levine","year":"2018","journal-title":"arXiv:1805.00909"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2008.4739438"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-012-5278-7"},{"key":"ref6","volume-title":"Abtive Inference, First InternationalWorkshop, IWAI 2020Co-Located With ECML\/PKDD 2020Ghent","author":"Verbelen","year":"2020"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmp.2017.09.004"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/s00422-019-00805-w"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/j.neuropsychologia.2018.09.013"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3389\/fncom.2018.00090"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/s00422-018-0753-2"},{"key":"ref12","first-page":"1255","article-title":"Modeling interaction via the principle of maximum causal entropy","volume-title":"Proc. 27th Int. Conf. Mach. Learn. (ICML)","author":"Ziebart"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.7551\/ecal_a_011"},{"key":"ref14","article-title":"Causal induction from visual observations for goal directed tasks","author":"Nair","year":"2019","journal-title":"arXiv:1910.01751"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/504729.504754"},{"key":"ref16","volume-title":"Reinforcement Learning and Optimal Control","author":"Bertsekas","year":"2019"},{"key":"ref17","first-page":"9","article-title":"Planning by probabilistic inference","volume-title":"Proc. 9th Int. Workshop Artif. Intell. Statist.","author":"Attias"},{"key":"ref18","volume-title":"Probabilistic Graphical Models: Principles and Techniques","author":"Koller","year":"2009"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/18.910573"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2004.1267047"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2015.2477379"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/s00500-021-05642-3"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2007.896497"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TAES.2015.140087"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-18164-6_17"},{"key":"ref26","first-page":"1961","article-title":"Point-based path prediction from polar histograms","volume-title":"Proc. 19th Int. Conf. Inf. Fusion (FUSION)","author":"Coscia"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.23919\/ICIF.2018.8455392"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/TAES.2018.2808098"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2017.11.006"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/MAES.2020.3032069"},{"key":"ref31","volume-title":"Reinforcement Learning: An Introduction","author":"Sutton","year":"2018"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511804779"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2009.5354147"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-64919-7_1"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-64919-7_2"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1609\/icaps.v23i1.13573"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.1974.1055186"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9668973\/09698186.pdf?arnumber=9698186","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,17]],"date-time":"2024-01-17T23:12:00Z","timestamp":1705533120000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9698186\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":37,"URL":"https:\/\/doi.org\/10.1109\/access.2022.3148127","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]}}}