{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T09:38:57Z","timestamp":1775209137966,"version":"3.50.1"},"reference-count":33,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"11","license":[{"start":{"date-parts":[[2023,11,1]],"date-time":"2023-11-01T00:00:00Z","timestamp":1698796800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,11,1]],"date-time":"2023-11-01T00:00:00Z","timestamp":1698796800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,11,1]],"date-time":"2023-11-01T00:00:00Z","timestamp":1698796800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Shenzhen Key Laboratory of Robotics and Computer Vision","award":["ZDSYS20220330160557001"],"award-info":[{"award-number":["ZDSYS20220330160557001"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2023,11]]},"DOI":"10.1109\/lra.2023.3320012","type":"journal-article","created":{"date-parts":[[2023,9,27]],"date-time":"2023-09-27T17:43:29Z","timestamp":1695836609000},"page":"7551-7558","source":"Crossref","is-referenced-by-count":81,"title":["GraspGPT: Leveraging Semantic Knowledge From a Large Language Model for Task-Oriented Grasping"],"prefix":"10.1109","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8287-7188","authenticated-orcid":false,"given":"Chao","family":"Tang","sequence":"first","affiliation":[{"name":"Shenzhen Key Laboratory of Robotics and Computer Vision, Southern University of Science and Technology, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-2693-9309","authenticated-orcid":false,"given":"Dehao","family":"Huang","sequence":"additional","affiliation":[{"name":"Shenzhen Key Laboratory of Robotics and Computer Vision, Southern University of Science and Technology, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-1925-0226","authenticated-orcid":false,"given":"Wenqi","family":"Ge","sequence":"additional","affiliation":[{"name":"Shenzhen Key Laboratory of Robotics and Computer Vision, Southern University of Science and Technology, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2113-5612","authenticated-orcid":false,"given":"Weiyu","family":"Liu","sequence":"additional","affiliation":[{"name":"Institute for Robotics and Intelligent Machines, Georgia Institute of Technology, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1677-6132","authenticated-orcid":false,"given":"Hong","family":"Zhang","sequence":"additional","affiliation":[{"name":"Shenzhen Key Laboratory of Robotics and Computer Vision, Southern University of Science and Technology, Shenzhen, China"}]}],"member":"263","reference":[{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2019.2933815"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9197289"},{"key":"ref15","first-page":"492","article-title":"LM-NAV: Robotic navigation with large pre-trained models of language, vision, and action","author":"shah","year":"0","journal-title":"Proc Conf Robot Learn"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-018-9784-8"},{"key":"ref31","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2014"},{"key":"ref30","doi-asserted-by":"crossref","DOI":"10.1109\/ICCV51070.2023.00371","article-title":"Segment anything","author":"kirillov","year":"2023"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2012.6385563"},{"key":"ref33","first-page":"1783","article-title":"TAX-Pose: Task-specific cross-pose estimation for robot manipulation","author":"pan","year":"0","journal-title":"Proc Conf Robot Learn"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160591"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9196971"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1177\/0278364919872545"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2020.2975706"},{"key":"ref17","doi-asserted-by":"crossref","DOI":"10.1109\/IROS55552.2023.10341577","article-title":"TidyBot: Personalized robot assistance with large language models","author":"wu","year":"2023"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160969"},{"key":"ref19","article-title":"VIMA: General robot manipulation with multimodal prompts","author":"jiang","year":"2022"},{"key":"ref18","first-page":"1769","article-title":"Inner monologue: Embodied reasoning through planning with language models","author":"huang","year":"0","journal-title":"Proc Conf Robot Learn PMLR"},{"key":"ref24","doi-asserted-by":"crossref","DOI":"10.1109\/IROS55552.2023.10342268","article-title":"Task-oriented grasp prediction with visual-language inputs","author":"tang","year":"2023"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1037\/0096-3445.104.3.192"},{"key":"ref26","first-page":"5105","article-title":"PointNet++ : Deep hierarchical feature learning on point sets in a metric space","author":"qi","year":"0","journal-title":"Proc 31st Int Conf Neural Inf Process Syst"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-021-10008-7"},{"key":"ref20","article-title":"Socratic models: Composing zero-shot multimodal reasoning with language","author":"zeng","year":"0","journal-title":"Proc 11th Int Conf Learn Representations"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48506.2021.9561877"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00299"},{"key":"ref28","first-page":"6000","article-title":"Attention is all you need","author":"vaswani","year":"0","journal-title":"Proc 31st Int Conf Neural Inf Process Syst"},{"key":"ref27","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"devlin","year":"2018"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1023\/B:BTTJ.0000047600.45421.6d"},{"key":"ref8","first-page":"1531","article-title":"Leveraging language for accelerated learning of tool manipulation","author":"ren","year":"0","journal-title":"Proc Conf Robot Learn"},{"key":"ref7","article-title":"Do as I can, not as I say: Grounding language in robotic affordances","author":"ahn","year":"2022"},{"key":"ref9","first-page":"9118","article-title":"Language models as zero-shot planners: Extracting actionable knowledge for embodied agents","author":"huang","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref4","first-page":"1540","article-title":"Same object, different grasps: Data and semantic knowledge for task-oriented grasping","author":"murali","year":"0","journal-title":"Proc Conf Robot Learn"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2010.5649406"},{"key":"ref6","article-title":"Lamda: Language models for dialog applications","author":"thoppilan","year":"2022"},{"key":"ref5","article-title":"Palm: Scaling language modeling with pathways","author":"chowdhery","year":"2022"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7083369\/10254630\/10265134.pdf?arnumber=10265134","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T09:57:08Z","timestamp":1730195828000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10265134\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11]]},"references-count":33,"journal-issue":{"issue":"11"},"URL":"https:\/\/doi.org\/10.1109\/lra.2023.3320012","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"value":"2377-3766","type":"electronic"},{"value":"2377-3774","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,11]]}}}