{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T04:09:19Z","timestamp":1774498159193,"version":"3.50.1"},"reference-count":55,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"am","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"NSF","award":["CCF-2112665"],"award-info":[{"award-number":["CCF-2112665"]}]},{"name":"NSF","award":["1730158"],"award-info":[{"award-number":["1730158"]}]},{"name":"NSF","award":["ACI-1541349"],"award-info":[{"award-number":["ACI-1541349"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2022,4]]},"DOI":"10.1109\/lra.2022.3144512","type":"journal-article","created":{"date-parts":[[2022,1,21]],"date-time":"2022-01-21T21:29:03Z","timestamp":1642800543000},"page":"3046-3053","source":"Crossref","is-referenced-by-count":44,"title":["Look Closer: Bridging Egocentric and Third-Person Views With Transformers for Robotic Manipulation"],"prefix":"10.1109","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3339-7513","authenticated-orcid":false,"given":"Rishabh","family":"Jangir","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9897-4003","authenticated-orcid":false,"given":"Nicklas","family":"Hansen","sequence":"additional","affiliation":[]},{"given":"Sambaran","family":"Ghosal","sequence":"additional","affiliation":[]},{"given":"Mohit","family":"Jain","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3150-778X","authenticated-orcid":false,"given":"Xiaolong","family":"Wang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1177\/0278364914549607"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1177\/0278364917710318"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2018.xiv.019"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2020.2988642"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992696"},{"key":"ref6","first-page":"1057","article-title":"Policy gradient methods for reinforcement learning with function approximation","volume-title":"Proc. Neural Inf. Process. Syst.","author":"Sutton","year":"1999"},{"key":"ref7","article-title":"Continuous control with deep reinforcement learning","volume-title":"Comput. Res. Repository","author":"Lillicrap","year":"2016"},{"key":"ref8","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Haarnoja","year":"2018"},{"issue":"1","key":"ref9","first-page":"1334","article-title":"End-to-end training of deep visuomotor policies","volume":"17","author":"Levine","year":"2016","journal-title":"J. Mach. Learn. Res."},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2016.7487517"},{"key":"ref11","first-page":"9191","article-title":"Visual reinforcement learning with imagined goals","volume-title":"Proc. Neural Inf. Process. Syst.","volume":"31","author":"Nair","year":"2018"},{"key":"ref12","article-title":"A framework for efficient robotic manipulation","volume-title":"arXiv:2012.07975","author":"Zhan","year":"2020"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989385"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794127"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1177\/0278364919887447"},{"key":"ref16","article-title":"Asymmetric self-play for automatic goal discovery in robotic manipulation","author":"OpenAI","year":"2021"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9561483"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1068\/p2935"},{"key":"ref19","first-page":"5998","article-title":"Attention is all you need","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Vaswani","year":"2017"},{"key":"ref20","article-title":"Image augmentation is all you need: Regularizing deep reinforcement learning from pixels","volume-title":"arXiv:2004.13649","author":"Kostrikov","year":"2021"},{"key":"ref21","article-title":"Learning predictive representations for deformable objects using contrastive estimation","volume-title":"arXiv:2003.05436","author":"Yan","year":"2020"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2018.xiv.008"},{"key":"ref23","first-page":"1282","article-title":"Quantifying generalization in reinforcement learning","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Cobbe","year":"2019"},{"key":"ref24","article-title":"Self-supervised policy adaptation during deployment","author":"Hansen","year":"2021"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01270-0_16"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00346"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8793485"},{"key":"ref28","article-title":"Learning vision-guided quadrupedal locomotion end-to-end with cross-modal transformers","author":"Yang","year":"2021","journal-title":"arXiv:2107.03996"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9197181"},{"key":"ref30","article-title":"Unsupervised learning of visual 3D keypoints for control","volume-title":"arXiv:2106.07643","author":"Chen","year":"2021"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/5.5968"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.1992.223215"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1177\/0278364911410755"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/s11554-013-0386-6"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2012.6385934"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01267-0_27"},{"key":"ref37","first-page":"422","article-title":"Reinforcement learning of active vision for manipulating objects under occlusions","volume-title":"Proc. Conf. Robot Learn.","author":"Cheng","year":"2018"},{"key":"ref38","article-title":"Active perception and representation for robotic manipulation","volume-title":"arXiv:2003.06734","author":"Zaky","year":"2020"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1810.04805"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref41","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2021"},{"key":"ref42","first-page":"7254","article-title":"Learning attentional communication for multi-agent cooperation","volume-title":"Proc. Neural Inf. Process. Syst.","volume":"31","author":"Jiang","year":"2020"},{"key":"ref43","article-title":"Decision transformer: Reinforcement learning via sequence modeling","volume-title":"arXiv:2106.01345","author":"Chen","year":"2021"},{"key":"ref44","article-title":"Reinforcement learning as one big sequence modeling problem","author":"Janner","year":"2021","journal-title":"arXiv:2106.02039"},{"key":"ref45","article-title":"Stabilizing deep q-learning with convnets and vision transformers under data augmentation","volume":"34","author":"Hansen","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref46","article-title":"Coarse-to-fine q-attention: Efficient learning for visual robotic manipulation via discretisation","volume-title":"arXiv:2106.12534","author":"James","year":"2021"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1016\/S0004-3702(98)00023-X"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/bf00115009"},{"key":"ref49","article-title":"Modeling purposeful adaptive behavior with the principle of maximum causal entropy","author":"Bagnell","year":"2010"},{"key":"ref50","first-page":"1433","article-title":"Maximum entropy inverse reinforcement learning","volume-title":"Proc. AAAI Conf. Artif. Intell.","author":"Ziebart","year":"2008"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2012.6386109"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17276"},{"key":"ref53","article-title":"Curl: Contrastive unsupervised representations for reinforcement learning","author":"Srinivas","year":"2020"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9561103"},{"key":"ref55","article-title":"Layer normalization","author":"Ba","year":"2016"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"https:\/\/ieeexplore.ieee.org\/ielam\/7083369\/9647862\/9690036-aam.pdf","content-type":"application\/pdf","content-version":"am","intended-application":"syndication"},{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7083369\/9647862\/09690036.pdf?arnumber=9690036","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,13]],"date-time":"2024-01-13T22:34:03Z","timestamp":1705185243000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9690036\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4]]},"references-count":55,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/lra.2022.3144512","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"value":"2377-3766","type":"electronic"},{"value":"2377-3774","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4]]}}}