{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T16:51:46Z","timestamp":1777654306594,"version":"3.51.4"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,5,13]],"date-time":"2024-05-13T00:00:00Z","timestamp":1715558400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,5,13]],"date-time":"2024-05-13T00:00:00Z","timestamp":1715558400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,5,13]]},"DOI":"10.1109\/icra57147.2024.10610779","type":"proceedings-article","created":{"date-parts":[[2024,8,8]],"date-time":"2024-08-08T17:51:05Z","timestamp":1723139465000},"page":"9086-9092","source":"Crossref","is-referenced-by-count":10,"title":["OpenAnnotate3D: Open-Vocabulary Auto-Labeling System for Multi-modal 3D Data"],"prefix":"10.1109","author":[{"given":"Yijie","family":"Zhou","sequence":"first","affiliation":[{"name":"Fudan University,China"}]},{"given":"Likun","family":"Cai","sequence":"additional","affiliation":[{"name":"University of Toronto,Canada"}]},{"given":"Xianhui","family":"Cheng","sequence":"additional","affiliation":[{"name":"Fudan University,China"}]},{"given":"Zhongxue","family":"Gan","sequence":"additional","affiliation":[{"name":"Fudan University,China"}]},{"given":"Xiangyang","family":"Xue","sequence":"additional","affiliation":[{"name":"Fudan University,China"}]},{"given":"Wenchao","family":"Ding","sequence":"additional","affiliation":[{"name":"Fudan University,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1177\/0278364913491297"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00939"},{"key":"ref5","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref6","article-title":"Gpt-4 technical report","year":"2023"},{"key":"ref7","article-title":"Palm: Scaling language modeling with pathways","author":"Chowdhery","year":"2022"},{"key":"ref8","article-title":"Emergent abilities of large language models","author":"Wei","year":"2022"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00085"},{"key":"ref10","article-title":"Tesla ai day 2022","year":"2000"},{"key":"ref11","article-title":"Chatgpt (september 17 version)","year":"2023"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-007-0090-8"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-012-0564-1"},{"key":"ref14","article-title":"Label Studio: Data labeling software","volume-title":"2020-2022, open source software","author":"Tkachenko"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3343031.3350535"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00071"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00096"},{"key":"ref18","article-title":"Computer vision annotation tool: A universal approach to data annotation. 2019","author":"Sekachev","year":"2019"},{"key":"ref19","article-title":"Roboflow (version 1.0)","volume-title":"computer Vision","author":"Dwyer","year":"2022"},{"key":"ref20","volume-title":"Labelbox","year":"2023"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2009.5457721"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/3DV.2017.00042"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160904"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC.2019.8916980"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-11024-6_39"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.23919\/MVA57639.2023.10216156"},{"key":"ref28","first-page":"28 492","article-title":"Robust speech recognition via large-scale weak super-vision","volume-title":"International Conference on Machine Learning","author":"Radford"},{"key":"ref29","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International conference on machine learning","author":"Radford"},{"key":"ref30","volume-title":"OpenAI","year":"2023"},{"key":"ref31","article-title":"Langchain","author":"Chase","year":"2022"},{"key":"ref32","article-title":"Grounding dino: Marrying dino with grounded pre-training for open-set object detection","author":"Liu","year":"2023"}],"event":{"name":"2024 IEEE International Conference on Robotics and Automation (ICRA)","location":"Yokohama, Japan","start":{"date-parts":[[2024,5,13]]},"end":{"date-parts":[[2024,5,17]]}},"container-title":["2024 IEEE International Conference on Robotics and Automation (ICRA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10609961\/10609862\/10610779.pdf?arnumber=10610779","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,10]],"date-time":"2024-08-10T05:55:25Z","timestamp":1723269325000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10610779\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,13]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/icra57147.2024.10610779","relation":{},"subject":[],"published":{"date-parts":[[2024,5,13]]}}}