{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T15:25:06Z","timestamp":1730301906686,"version":"3.28.0"},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,6,24]],"date-time":"2024-06-24T00:00:00Z","timestamp":1719187200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,6,24]],"date-time":"2024-06-24T00:00:00Z","timestamp":1719187200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,6,24]]},"DOI":"10.1109\/ur61395.2024.10597480","type":"proceedings-article","created":{"date-parts":[[2024,7,26]],"date-time":"2024-07-26T17:23:58Z","timestamp":1722014638000},"page":"273-280","source":"Crossref","is-referenced-by-count":0,"title":["Curiosity-Driven Learning for Visual Control of Nonholonomic Mobile Robots"],"prefix":"10.1109","author":[{"given":"Takieddine","family":"Soualhi","sequence":"first","affiliation":[{"name":"Belfort-Montb&#x00E9;liard University of Technology, UTBM, CIAD,Belfort,France,90010"}]},{"given":"Nathan","family":"Crombez","sequence":"additional","affiliation":[{"name":"Belfort-Montb&#x00E9;liard University of Technology, UTBM, CIAD,Belfort,France,90010"}]},{"given":"Alexandre","family":"Lombard","sequence":"additional","affiliation":[{"name":"Belfort-Montb&#x00E9;liard University of Technology, UTBM, CIAD,Belfort,France,90010"}]},{"given":"St\u00e9phane","family":"Galland","sequence":"additional","affiliation":[{"name":"Belfort-Montb&#x00E9;liard University of Technology, UTBM, CIAD,Belfort,France,90010"}]},{"given":"Yassine","family":"Ruichek","sequence":"additional","affiliation":[{"name":"Belfort-Montb&#x00E9;liard University of Technology, UTBM, CIAD,Belfort,France,90010"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.1999.827879"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/MRA.2006.250573"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.2007.363646"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2011.2162765"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.1994.407471"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1016\/j.robot.2014.03.007"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/URAI.2015.7358936"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.5555\/2999134.2999257"},{"key":"ref9","article-title":"Sim2real view invariant visual servoing by recurrent control","author":"Sadeghi","year":"2017","journal-title":"arXiv preprint"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2019.XV.055"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CASE48305.2020.9216986"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9197196"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989381"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ECMR.2019.8870964"},{"key":"ref15","article-title":"Memory-augmented reinforcement learning for image-goal navigation","author":"Mezghani","year":"2021","journal-title":"arXiv preprint"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00945"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/UR52253.2021.9494664"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9197136"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2006.890271"},{"key":"ref20","article-title":"# exploration: A study of count-based exploration for deep reinforcement learning","volume":"30","author":"Tang","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref21","article-title":"Exploration by random network distillation","author":"Burda","year":"2018","journal-title":"arXiv preprint"},{"key":"ref22","article-title":"Never give up: Learning directed exploration strategies","author":"Badia","year":"2020","journal-title":"arXiv preprint"},{"key":"ref23","article-title":"Rewarding episodic visitation discrepancy for exploration in reinforcement learning","author":"Yuan","year":"2022","journal-title":"arXiv preprint"},{"key":"ref24","article-title":"Deep recurrent q-learning for partially observable mdps","volume-title":"2015 aaai fall symposium series","author":"Hausknecht","year":"2015"},{"key":"ref25","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017","journal-title":"arXiv preprint"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref27","article-title":"Large-scale study of curiosity-driven learning","author":"Burda","year":"2018","journal-title":"arXiv preprint"},{"issue":"1","key":"ref28","first-page":"12348","article-title":"Stable-baselines3: Reliable reinforcement learning implementations","volume":"22","author":"Raffin","year":"2021","journal-title":"The Journal of Machine Learning Research"}],"event":{"name":"2024 21st International Conference on Ubiquitous Robots (UR)","start":{"date-parts":[[2024,6,24]]},"location":"New York, NY, USA","end":{"date-parts":[[2024,6,27]]}},"container-title":["2024 21st International Conference on Ubiquitous Robots (UR)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10597436\/10597437\/10597480.pdf?arnumber=10597480","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,27]],"date-time":"2024-07-27T05:16:59Z","timestamp":1722057419000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10597480\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,24]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/ur61395.2024.10597480","relation":{},"subject":[],"published":{"date-parts":[[2024,6,24]]}}}