{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T20:10:13Z","timestamp":1764015013400,"version":"3.45.0"},"reference-count":61,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1109\/lra.2025.3623027","type":"journal-article","created":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T17:42:36Z","timestamp":1760722956000},"page":"12955-12962","source":"Crossref","is-referenced-by-count":0,"title":["Don\u2019t Let Your Robot Be Harmful: Responsible Robotic Manipulation via Safety-As-Policy"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6483-0650","authenticated-orcid":false,"given":"Minheng","family":"Ni","sequence":"first","affiliation":[{"name":"Hong Kong Polytechnic University, Kowloon, Hong Kong, SAR, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6302-8348","authenticated-orcid":false,"given":"Lei","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Hamburg, Hamburg, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8672-5816","authenticated-orcid":false,"given":"Zihan","family":"Chen","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Harbin, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8579-0547","authenticated-orcid":false,"given":"Kaixin","family":"Bai","sequence":"additional","affiliation":[{"name":"University of Hamburg, Hamburg, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0375-9146","authenticated-orcid":false,"given":"Zhaopeng","family":"Chen","sequence":"additional","affiliation":[{"name":"Agile Robots, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7856-5760","authenticated-orcid":false,"given":"Jianwei","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Hamburg, Hamburg, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2078-4215","authenticated-orcid":false,"given":"Lei","family":"Zhang","sequence":"additional","affiliation":[{"name":"Hong Kong Polytechnic University, Kowloon, Hong Kong, SAR, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3330-783X","authenticated-orcid":false,"given":"Wangmeng","family":"Zuo","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Harbin, China"}]}],"member":"263","reference":[{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1109\/JAS.2017.7510604"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.17973\/MMSJ.2016_06_201611"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/ACCESS.2023.3310935"},{"key":"ref4","first-page":"69900","article-title":"STEVE-1: A generative model for text-to-behavior in minecraft","volume-title":"Proc. 37th Adv. Conf. Neural Inf. Process. Syst.","author":"Lifshitz","year":"2024"},{"year":"2024","author":"Zhao","article-title":"Steve series: Step-by-step construction of agent systems in minecraft","key":"ref5"},{"year":"2024","author":"Zhang","article-title":"Contactdexnet: Multi-fingered robotic hand grasping in cluttered environments through hand-object contact semantic mapping","key":"ref6"},{"year":"2025","author":"Dong","article-title":"M4diffuser: Multi-view diffusion policy with manipulability-aware control for robust mobile manipulation","key":"ref7"},{"year":"2025","author":"Xu","article-title":"Funcanon: Learning pose-aware action primitives via functional object canonicalization for generalizable robotic manipulation","key":"ref8"},{"doi-asserted-by":"publisher","key":"ref9","DOI":"10.1109\/IROS51168.2021.9636342"},{"year":"2023","author":"Xu","article-title":"Creative robot tool use with large language models","key":"ref10"},{"year":"2023","author":"Zhou","article-title":"Generalizable long-horizon manipulations with large language models","key":"ref11"},{"doi-asserted-by":"publisher","key":"ref12","DOI":"10.15607\/rss.2023.xix.025"},{"year":"2023","author":"Yu","article-title":"Language to rewards for robotic skill synthesis","key":"ref13"},{"year":"2023","author":"Ma","article-title":"Eureka: Human-level reward design via coding large language models","key":"ref14"},{"doi-asserted-by":"publisher","key":"ref15","DOI":"10.1146\/annurev-control-060117-104848"},{"doi-asserted-by":"publisher","key":"ref16","DOI":"10.1126\/science.aat8414"},{"year":"2022","author":"Nair","article-title":"R3m: A universal visual representation for robot manipulation","key":"ref17"},{"doi-asserted-by":"publisher","key":"ref18","DOI":"10.52202\/079017-3188"},{"doi-asserted-by":"publisher","key":"ref19","DOI":"10.4172\/2168-9695.1000139"},{"doi-asserted-by":"publisher","key":"ref20","DOI":"10.1016\/j.inffus.2019.12.012"},{"key":"ref21","first-page":"80079","article-title":"Jailbroken: How does LLM safety training fail?","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Wei","year":"2024"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.1109\/CVPR52729.2023.02157"},{"doi-asserted-by":"publisher","key":"ref23","DOI":"10.1109\/CVPR52688.2022.01042"},{"issue":"2","key":"ref24","first-page":"275","article-title":"Advancing nsfw detection in AI: Training models to detect drawings, animations, and assess degrees of sexiness","volume":"2","author":"Guzman","year":"2023","journal-title":"J. Knowl. Learn. Sci. Technol."},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.1109\/ICSE.2012.6227177"},{"year":"2023","author":"Achiam","article-title":"GPT-4 technical report","key":"ref26"},{"year":"2023","author":"Team","article-title":"Gemini: A family of highly capable multimodal models","key":"ref27"},{"year":"2024","author":"Dubey","article-title":"The LLAMA 3 herd of models","key":"ref28"},{"year":"2022","author":"Bai","article-title":"Training a helpful and harmless assistant with reinforcement learning from human feedback","key":"ref29"},{"doi-asserted-by":"publisher","key":"ref30","DOI":"10.1109\/SP40001.2021.00019"},{"doi-asserted-by":"publisher","key":"ref31","DOI":"10.1038\/s42256-025-00985-0"},{"doi-asserted-by":"publisher","key":"ref32","DOI":"10.18653\/v1\/2024.findings-acl.107"},{"year":"2023","author":"Pawelczyk","article-title":"In-context unlearning: Language models as few shot unlearners","key":"ref33"},{"year":"2024","author":"Naseh","article-title":"Injecting bias in text-to-image models via composite-trigger backdoors","key":"ref34"},{"doi-asserted-by":"publisher","key":"ref35","DOI":"10.1109\/CVPR52733.2024.01141"},{"doi-asserted-by":"publisher","key":"ref36","DOI":"10.1007\/978-3-031-73661-2_20"},{"doi-asserted-by":"publisher","key":"ref37","DOI":"10.1609\/aaai.v38i19.30144"},{"doi-asserted-by":"publisher","key":"ref38","DOI":"10.1007\/978-3-031-72670-5_18"},{"doi-asserted-by":"publisher","key":"ref39","DOI":"10.1109\/IROS45743.2020.9341522"},{"doi-asserted-by":"publisher","key":"ref40","DOI":"10.1146\/annurev-control-071723-102940"},{"key":"ref41","first-page":"25621","article-title":"Learning barrier certificates: Towards safe reinforcement learning with zero training-time violations","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Luo","year":"2021"},{"key":"ref42","first-page":"1877","article-title":"Learning to be safe: Deep RL with a safety critic","volume":"33","author":"Srinivasan","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"doi-asserted-by":"publisher","key":"ref43","DOI":"10.1109\/LRA.2024.3396368"},{"key":"ref44","first-page":"1554","article-title":"Safe driving via expert guided policy optimization","volume-title":"Proc. Conf. Robot Learn.","author":"Peng","year":"2022"},{"doi-asserted-by":"publisher","key":"ref45","DOI":"10.15607\/rss.2025.xxi.010"},{"year":"2024","author":"Kim","article-title":"Openvla: An open-source vision-language-action model","key":"ref46"},{"year":"2023","author":"Brohan","article-title":"RT-2: Vision-language-action models transfer web knowledge to robotic control","key":"ref47"},{"year":"2023","author":"Li","article-title":"Vision-language foundation models as effective robot imitators","key":"ref48"},{"key":"ref49","first-page":"287","article-title":"Do as i can, not as i say: Grounding language in robotic affordances","volume-title":"Proc. Conf. Robot Learn.","author":"Brohan","year":"2023"},{"doi-asserted-by":"publisher","key":"ref50","DOI":"10.1109\/ICRA48891.2023.10160591"},{"doi-asserted-by":"publisher","key":"ref51","DOI":"10.1007\/s10514-023-10139-z"},{"doi-asserted-by":"publisher","key":"ref52","DOI":"10.1007\/s10514-023-10131-7"},{"key":"ref53","first-page":"540","article-title":"Voxposer: Composable 3D value maps for robotic manipulation with language models","volume-title":"Proc. 7th Annu. Conf. Robot Learn.","author":"Huang","year":"2023"},{"doi-asserted-by":"publisher","key":"ref54","DOI":"10.1109\/LRA.2024.3477090"},{"year":"2024","author":"Li","article-title":"What foundation models can bring for robot learning in manipulation: A survey","key":"ref55"},{"year":"2023","author":"Zhou","article-title":"Language-conditioned learning for robotic manipulation: A survey","key":"ref56"},{"year":"2024","author":"Hurst","article-title":"Gpt-4o system card","key":"ref57"},{"key":"ref58","first-page":"8821","article-title":"Zero-shot text-to-image generation","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Ramesh","year":"2021"},{"year":"2020","author":"Brown","article-title":"Language models are few-shot learners","key":"ref59"},{"year":"2024","author":"Ni","article-title":"Visual-o1: Understanding ambiguous instructions via multi-modal multi-turn chain-of-thoughts reasoning","key":"ref60"},{"key":"ref61","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Wei","year":"2022"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/7083369\/11215960\/11206481.pdf?arnumber=11206481","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T19:02:59Z","timestamp":1764010979000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11206481\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":61,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/lra.2025.3623027","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"type":"electronic","value":"2377-3766"},{"type":"electronic","value":"2377-3774"}],"subject":[],"published":{"date-parts":[[2025,12]]}}}