{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T12:50:31Z","timestamp":1774702231804,"version":"3.50.1"},"reference-count":36,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,8,1]],"date-time":"2019-08-01T00:00:00Z","timestamp":1564617600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,8,1]],"date-time":"2019-08-01T00:00:00Z","timestamp":1564617600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,8,1]],"date-time":"2019-08-01T00:00:00Z","timestamp":1564617600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,8]]},"DOI":"10.1109\/devlrn.2019.8850723","type":"proceedings-article","created":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T03:40:48Z","timestamp":1569901248000},"page":"1-7","source":"Crossref","is-referenced-by-count":16,"title":["Efficient Intrinsically Motivated Robotic Grasping with Learning-Adaptive Imagination in Latent Space"],"prefix":"10.1109","author":[{"given":"Muhammad Burhan","family":"Hafez","sequence":"first","affiliation":[]},{"given":"Cornelius","family":"Weber","sequence":"additional","affiliation":[]},{"given":"Matthias","family":"Kerzel","sequence":"additional","affiliation":[]},{"given":"Stefan","family":"Wermter","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.1999.831553"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-27645-3_7"},{"key":"ref31","article-title":"Continuous control with deep reinforcement learning","author":"lillicrap","year":"0","journal-title":"Proceedings of the International Conference on Learning Representations (ICLR)"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1016\/B978-1-55860-141-3.50030-4"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.2017.8172289"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2013.6696520"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/DEVLRN.2017.8329785"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.tics.2013.09.001"},{"key":"ref11","first-page":"509","article-title":"Slowness-based neural visuomotor control with an intrinsically motivated continuous actor-critic","author":"hafez","year":"0","journal-title":"Proceedings of the 26th European Symposium on Artificial Neural Networks Computational Intelligence and Machine Learning (ESANN)"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1515\/pjbr-2019-0005"},{"key":"ref13","author":"forestier","year":"2017","journal-title":"Intrinsically motivated goal exploration processes with automatic curriculum learning"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.3389\/fnbot.2018.00030"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/SMC.2014.6974211"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-015-1861-8"},{"key":"ref17","author":"ha","year":"2018","journal-title":"World Models"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1098\/rstb.2008.0314"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1016\/j.neuropsychologia.2015.04.005"},{"key":"ref28","author":"feinberg","year":"2018","journal-title":"Model-Based Value Expansion for Efficient Model-Free Reinforcement Learning"},{"key":"ref4","first-page":"5048","article-title":"Hindsight experience replay","author":"andrychowicz","year":"2017","journal-title":"Advances in Neural Information Systems (NIPS)"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8463189"},{"key":"ref3","article-title":"Prioritized experience replay","author":"schaul","year":"0","journal-title":"Proceedings of the International Conference on Learning Representations (ICLR)"},{"key":"ref6","first-page":"2750","article-title":"# Exploration: A study of count-based exploration for deep reinforcement learning","author":"tang","year":"2017","journal-title":"Advances in Neural Information Processing Systems (NIPS)"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2019.8852254"},{"key":"ref5","article-title":"Hierarchical reinforcement learning with hindsight","author":"levy","year":"0","journal-title":"Proc of the Int Conf on Learning Representations (ICLR)"},{"key":"ref8","author":"stadie","year":"2015","journal-title":"Incentivizing exploration in reinforcement rearning with deep predictive models"},{"key":"ref7","author":"machado","year":"2019","journal-title":"Count-based exploration with the successor representation"},{"key":"ref2","first-page":"1334","article-title":"End-to-end training of deep visuomotor policies","volume":"17","author":"levine","year":"2016","journal-title":"Journal of Machine Learning Research"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2017.70"},{"key":"ref1","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","article-title":"Humanlevel control through deep reinforcement learning","volume":"518","author":"mnih","year":"2015","journal-title":"Nature"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1016\/j.tics.2017.05.002"},{"key":"ref22","first-page":"7988","article-title":"Imagination machines: A new challenge for artificial intelligence","author":"mahadevan","year":"0","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1037\/0021-9010.79.4.481"},{"key":"ref24","first-page":"2829","article-title":"Continuous deep Q-learning with model-based acceleration","author":"gu","year":"0","journal-title":"Proceedings of the International Conference on Machine Learning (ICML)"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.cobeha.2018.12.011"},{"key":"ref26","first-page":"5690","article-title":"Imagination-augmented agents for deep reinforcement learning","author":"racanire","year":"2017","journal-title":"Advances in Neural Information Processing Systems (NIPS)"},{"key":"ref25","first-page":"195","article-title":"Uncertainty-driven imagination for continuous deep reinforcement learning","author":"kalweit","year":"0","journal-title":"Proceedings of the 1st Annual Conference on Robot Learning"}],"event":{"name":"2019 Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)","location":"Oslo, Norway","start":{"date-parts":[[2019,8,19]]},"end":{"date-parts":[[2019,8,22]]}},"container-title":["2019 Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8843356\/8850675\/08850723.pdf?arnumber=8850723","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,17]],"date-time":"2022-07-17T21:49:40Z","timestamp":1658094580000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8850723\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,8]]},"references-count":36,"URL":"https:\/\/doi.org\/10.1109\/devlrn.2019.8850723","relation":{},"subject":[],"published":{"date-parts":[[2019,8]]}}}