{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T10:51:44Z","timestamp":1766055104103,"version":"3.48.0"},"reference-count":50,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"am","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Berkeley AI Research at the University of California, Berkeley and Toyota Motor North America","award":["W911NF-17-2-0181"],"award-info":[{"award-number":["W911NF-17-2-0181"]}]},{"name":"NSF","award":["IIS-2246811"],"award-info":[{"award-number":["IIS-2246811"]}]},{"name":"NSF","award":["IIS-2150826"],"award-info":[{"award-number":["IIS-2150826"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1109\/lra.2025.3640424","type":"journal-article","created":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T18:38:46Z","timestamp":1764873526000},"page":"1242-1249","source":"Crossref","is-referenced-by-count":0,"title":["Learning to Drive Anywhere With Model-Based Reannotation"],"prefix":"10.1109","volume":"11","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0361-7383","authenticated-orcid":false,"given":"Noriaki","family":"Hirose","sequence":"first","affiliation":[{"name":"Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA, USA"}]},{"given":"Lydia","family":"Ignatova","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9880-7261","authenticated-orcid":false,"given":"Kyle","family":"Stachowicz","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA, USA"}]},{"given":"Catherine","family":"Glossop","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6764-2743","authenticated-orcid":false,"given":"Sergey","family":"Levine","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7541-3278","authenticated-orcid":false,"given":"Dhruv","family":"Shah","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA, USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref2","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"33","author":"Brown","year":"2020"},{"key":"ref3","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford","year":"2021"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"ref5","first-page":"1","article-title":"Dinov2: Learning robust visual features without supervision","author":"Oquab","year":"2024","journal-title":"Trans. Mach. Learn. Res. J."},{"key":"ref6","first-page":"6892","article-title":"Open X-embodiment: Robotic learning datasets and RT-X models","volume-title":"Proc. IEEE Int. Conf. Robot. Automat.","author":"ONeill","year":"2024"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.120"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10161227"},{"key":"ref9","first-page":"711","article-title":"ViNT: A foundation model for visual navigation","volume-title":"Proc. Conf. Robot Learn.","author":"Shah","year":"2023"},{"key":"ref10","article-title":"Semi-parametric topological memory for navigation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Savinov","year":"2018"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00278"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2019.2925731"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9196644"},{"key":"ref14","first-page":"674","article-title":"Rapid exploration for open-world navigation with latent goal models","volume-title":"Proc. Conf. Robot Learn.","author":"Shah","year":"2022"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2022.XVIII.019"},{"key":"ref16","first-page":"3100","article-title":"FastRLAP: A system for learning high-speed driving via deep RL and autonomous practicing","volume-title":"Proc. Conf. Robot Learn.","author":"Stachowicz","year":"2023"},{"key":"ref17","article-title":"SELFI: Autonomous self-improvement with RL for vision-based navigation around people","volume-title":"Proc. Conf. Robot Learn.","author":"Hirose","year":"2024"},{"key":"ref18","article-title":"Lifelong autonomous fine-tuning for navigation foundation models","volume-title":"Proc. Conf. Robot Learn.","author":"Stachowicz","year":"2024"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160761"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2018.8594031"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52734.2025.00645"},{"key":"ref22","first-page":"24639","article-title":"Video pretraining (VPT): Learning to act by watching unlabeled online videos","volume-title":"Proc. Neural Inf. Process. Syst.","volume":"35","author":"Baker","year":"2022"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2021.3075644"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52734.2025.01556"},{"key":"ref25","first-page":"16558","article-title":"DROID-SLAM: Deep visual SLAM for monocular, stereo, and RGB-D cameras","volume-title":"Proc. Adv. Neural Inf. Process.","author":"Teed","year":"2011"},{"key":"ref26","first-page":"39033","article-title":"Deep patch visual odometry","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Teed","year":"2023"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA57147.2024.10610665"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/IROS47612.2022.9981751"},{"key":"ref29","article-title":"LeLAN: Learning a language-conditioned navigation policy from in-the-wild video","volume-title":"Proc. Conf. Robot Learn.","author":"Hirose","year":"2024"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1177\/0278364915614638"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1007\/s11370-020-00324-9"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3070543"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3184025"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA46639.2022.9811648"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2023.3329626"},{"year":"2025","key":"ref37","article-title":"FrodoBots-2 k"},{"year":"2025","key":"ref38","article-title":"Earth rovers SDK"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8460655"},{"key":"ref40","first-page":"619","article-title":"Semantic terrain classification for off-road autonomous driving","volume-title":"Proc. Conf. Robot Learn.","author":"Shaban","year":"2022"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1115\/1.3662552"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TIE.2011.2162714"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2023.3273786"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1016\/j.dche.2023.100084"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1016\/j.automatica.2004.04.003"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1016\/S0959-1524(02)00090-2"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.045"},{"year":"2025","key":"ref48","article-title":"Intel RealSense T265"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/IROS47612.2022.9981958"},{"article-title":"OmniVLA: An omni-modal vision-language-action model for robot navigation","year":"2025","author":"Hirose","key":"ref50"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"https:\/\/ieeexplore.ieee.org\/ielam\/7083369\/11293803\/11278012-aam.pdf","content-type":"application\/pdf","content-version":"am","intended-application":"syndication"},{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/7083369\/11293803\/11278012.pdf?arnumber=11278012","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T10:48:39Z","timestamp":1766054919000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11278012\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":50,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/lra.2025.3640424","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"type":"electronic","value":"2377-3766"},{"type":"electronic","value":"2377-3774"}],"subject":[],"published":{"date-parts":[[2026,2]]}}}