{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,28]],"date-time":"2025-10-28T05:57:05Z","timestamp":1761631025654,"version":"3.37.3"},"reference-count":71,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"6","license":[{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100015539","name":"Australian Government","doi-asserted-by":"publisher","award":["AUSMURIB000001"],"award-info":[{"award-number":["AUSMURIB000001"]}],"id":[{"id":"10.13039\/100015539","id-type":"DOI","asserted-by":"publisher"}]},{"name":"ONR MURI","award":["N00014-19-1-2571"],"award-info":[{"award-number":["N00014-19-1-2571"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Automat. Contr."],"published-print":{"date-parts":[[2023,6]]},"DOI":"10.1109\/tac.2023.3250159","type":"journal-article","created":{"date-parts":[[2023,2,28]],"date-time":"2023-02-28T18:57:48Z","timestamp":1677610668000},"page":"3557-3572","source":"Crossref","is-referenced-by-count":12,"title":["Smoother Entropy for Active State Trajectory Estimation and Obfuscation in POMDPs"],"prefix":"10.1109","volume":"68","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6797-5617","authenticated-orcid":false,"given":"Timothy L.","family":"Molloy","sequence":"first","affiliation":[{"name":"Department of Electrical and Electronic Engineering, University of Melbourne, VIC, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4342-209X","authenticated-orcid":false,"given":"Girish N.","family":"Nair","sequence":"additional","affiliation":[{"name":"Department of Electrical and Electronic Engineering, University of Melbourne, Parkville, VIC, Australia"}]}],"member":"263","reference":[{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ALLERTON.2018.8636086"},{"journal-title":"H&#x00F6;lder and Locally H&#x00F6;lder Continuous Functions and Open Sets of Class Ck Ck ?","year":"2017","author":"fiorenza","key":"ref57"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2009.2034206"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511976155"},{"journal-title":"Probabilistic Robotics","year":"2005","author":"thrun","key":"ref15"},{"key":"ref59","doi-asserted-by":"crossref","first-page":"33","DOI":"10.1613\/jair.678","article-title":"Value-function approximations for partially observable Markov decision processes","volume":"13","author":"hauskrecht","year":"2000","journal-title":"J Artif Intell Res"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2016.7799127"},{"key":"ref58","volume":"2","author":"bertsekas","year":"2012","journal-title":"Dynamic Programming and Optimal Control"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2005.850223"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-60603-3_5"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2019.2948598"},{"key":"ref55","first-page":"131","article-title":"Approximate inference using conditional entropy decompositions","author":"globerson","year":"0","journal-title":"Proc Int Conf Artif Intell Statist"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/GlobalSIP.2018.8646408"},{"key":"ref54","volume":"1","author":"bertsekas","year":"1995","journal-title":"Dynamic Programming and Optimal Control"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2005.I.009"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.1999.769927"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2005.1570193"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2012.6385637"},{"key":"ref51","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","author":"haarnoja","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref50","first-page":"1352","article-title":"Reinforcement learning with deep energy-based policies","author":"haarnoja","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"journal-title":"Elements of Information Theory","year":"2006","author":"cover","key":"ref46"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT.2005.1523313"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/s10463-009-0236-2"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.1974.1100635"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TCOM.1973.1091610"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/18.272494"},{"journal-title":"Directed Information for Channels with Feedback","year":"1998","author":"kramer","key":"ref44"},{"key":"ref43","first-page":"303","article-title":"Causality, feedback and directed information","author":"massey","year":"0","journal-title":"Proc Int Symp Inf Theory Appl"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2021.XVII.041"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2014.2362098"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9781316471104"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2017.2664049"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/S0005-1098(04)00202-X"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2008.2006100"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TNN.2007.899251"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2007.01.006"},{"key":"ref40","first-page":"1","article-title":"$\\rho$-POMDPs have Lipschitz-continuous epsilon-optimal value functions","author":"fehr","year":"0","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1002\/0471221279"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2017.8263874"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/ACC.2015.7170902"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2015.7402921"},{"key":"ref31","first-page":"50","article-title":"DESPOT-Alpha: Online POMDP planning with large state and observation spaces","author":"garg","year":"0","journal-title":"Proc Robot Sci Syst"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2008.IV.009"},{"key":"ref33","first-page":"64","article-title":"A POMDP extension with belief-dependent rewards","author":"araya","year":"0","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2007.897908"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.23919\/ECC54610.2021.9654854"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.23919\/ACC50511.2021.9483142"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2021.3055022"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1016\/j.arcontrol.2019.04.006"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2020.3019177"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2019.2956676"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2019.2922583"},{"key":"ref68","first-page":"1043","article-title":"Coastal navigation with mobile robots","author":"roy","year":"0","journal-title":"Proc 12th Int Conf Neural Inf Process Syst"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.23919\/ACC.2017.7963192"},{"key":"ref67","first-page":"2378","article-title":"Monte Carlo information-oriented planning","author":"thomas","year":"0","journal-title":"Proc 24th Eur Conf Artif Intell"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/RAMECH.2011.6070460"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2020.3005634"},{"key":"ref69","first-page":"1785","article-title":"How to protect your privacy? A framework for counter-adversarial decision making","author":"louren\u00e7o","year":"0","journal-title":"Proc IEEE 59th Conf Decis Control"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2018.2809005"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1017\/S0263574706002931"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2008.2006706"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-15-0493-8"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1016\/B978-1-55860-377-6.50052-9"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2020.109278"},{"key":"ref65","first-page":"1088","article-title":"Approximating optimal policies for partially observable stochastic domains","volume":"2","author":"parr","year":"0","journal-title":"Proc 14th Int Joint Conf Artif Intell"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2019.2942528"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CDC40024.2019.9030012"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1.11324"},{"key":"ref60","first-page":"7643","article-title":"An on-line POMDP solver for continuous observation spaces","author":"hoerger","year":"0","journal-title":"Proc IEEE Int Conf Robot Autom"},{"journal-title":"Hidden Markov Models Estimation and Control","year":"1995","author":"elliott","key":"ref62"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794121"}],"container-title":["IEEE Transactions on Automatic Control"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9\/10137341\/10054483.pdf?arnumber=10054483","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,19]],"date-time":"2023-06-19T18:30:52Z","timestamp":1687199452000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10054483\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6]]},"references-count":71,"journal-issue":{"issue":"6"},"URL":"https:\/\/doi.org\/10.1109\/tac.2023.3250159","relation":{},"ISSN":["0018-9286","1558-2523","2334-3303"],"issn-type":[{"type":"print","value":"0018-9286"},{"type":"electronic","value":"1558-2523"},{"type":"electronic","value":"2334-3303"}],"subject":[],"published":{"date-parts":[[2023,6]]}}}