{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,25]],"date-time":"2026-01-25T16:56:23Z","timestamp":1769360183315,"version":"3.49.0"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,11,26]],"date-time":"2024-11-26T00:00:00Z","timestamp":1732579200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,11,26]],"date-time":"2024-11-26T00:00:00Z","timestamp":1732579200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,11,26]]},"DOI":"10.1109\/fllm63129.2024.10852462","type":"proceedings-article","created":{"date-parts":[[2025,1,28]],"date-time":"2025-01-28T18:35:23Z","timestamp":1738089323000},"page":"91-96","source":"Crossref","is-referenced-by-count":2,"title":["Robots can feel: LLM-based Framework for Robot Ethical Reasoning"],"prefix":"10.1109","author":[{"given":"Artem","family":"Lykov","sequence":"first","affiliation":[{"name":"Skolkovo Institute of Science and Technology,Intelligent Space Robotics Laboratory,Moscow,Russia"}]},{"given":"Miguel Altamirano","family":"Cabrera","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology,Intelligent Space Robotics Laboratory,Moscow,Russia"}]},{"given":"Koffivi","family":"Fid\u00e8le Gbagbe","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology,Intelligent Space Robotics Laboratory,Moscow,Russia"}]},{"given":"Dzmitry","family":"Tsetserukou","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology,Intelligent Space Robotics Laboratory,Moscow,Russia"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Introducing ChatGPT","year":"2022"},{"key":"ref2","article-title":"PaLM-E: an embodied multimodal language model","volume-title":"Proceedings of the 40th International Conference on Machine Learning, ICML\u201923","author":"Driess"},{"key":"ref3","article-title":"Rt-2: Vision-language-action models transfer web knowledge to robotic control","author":"Brohan","year":"2023"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.54097\/hset.v39i.6767"},{"key":"ref5","article-title":"Agility\u2019s Latest Digit Robot Prepares for its First Job","volume-title":"IEEE Spectrum","author":"Ackerman"},{"key":"ref6","first-page":"712","article-title":"Cognitive-Dog: Large Multimodal Model Based System to Translate Vision and Language into Action of Quadruped Robot","volume-title":"Companion of the 2024 ACM\/IEEE International Conference on Human-Robot Interaction, HRI \u201924","author":"Lykov"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.birob.2023.100131"},{"key":"ref8","article-title":"A study on improving reasoning in language models","volume-title":"I Can\u2019t Believe It\u2019s Not Better Workshop: Failure Modes in the Age of Foundation Models","author":"Du"},{"key":"ref9","volume-title":"The Bicentennial Man and Other Stories","author":"Asimov","year":"1976"},{"key":"ref10","article-title":"Cognitiveos: Large multimodal model based system to endow any type of robot with generative ai","author":"Lykov","year":"2024"},{"key":"ref11","article-title":"Autort: Embodied foundation models for large scale orchestration of robotic agents","author":"Ahn","year":"2024"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.4324\/9781315719696-8"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-021-00778-6"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.3389\/frobt.2023.1271610"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.726"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-024-53255-1"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.2307\/2233985"},{"key":"ref18","article-title":"Gpt-4 technical report","year":"2023"},{"key":"ref19","article-title":"Chatbot arena: An open platform for evaluating llms by human preference","author":"Chiang","year":"2024"},{"key":"ref20","article-title":"Introducing Gemini: Google\u2019s most capable AI model yet","author":"Pichai","year":"2023"},{"key":"ref21","article-title":"Introducing the next generation of Claude","volume-title":"Anthropic News","year":"2024"},{"key":"ref22","article-title":"Llama 3 model card","year":"2024"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-55560-2_5"},{"key":"ref24","article-title":"Library for accessing GigaChat","year":"2024"},{"key":"ref25","article-title":"Qwen technical report","author":"Bai","year":"2023"}],"event":{"name":"2024 2nd International Conference on Foundation and Large Language Models (FLLM)","location":"Dubai, United Arab Emirates","start":{"date-parts":[[2024,11,26]]},"end":{"date-parts":[[2024,11,29]]}},"container-title":["2024 2nd International Conference on Foundation and Large Language Models (FLLM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10852419\/10852420\/10852462.pdf?arnumber=10852462","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,29]],"date-time":"2025-01-29T18:45:31Z","timestamp":1738176331000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10852462\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,26]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/fllm63129.2024.10852462","relation":{},"subject":[],"published":{"date-parts":[[2024,11,26]]}}}