{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T14:30:49Z","timestamp":1776090649760,"version":"3.50.1"},"reference-count":20,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,8,26]],"date-time":"2024-08-26T00:00:00Z","timestamp":1724630400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,8,26]],"date-time":"2024-08-26T00:00:00Z","timestamp":1724630400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006180","name":"Technology Development","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006180","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,8,26]]},"DOI":"10.1109\/ro-man60168.2024.10731235","type":"proceedings-article","created":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T17:45:45Z","timestamp":1730310345000},"page":"224-230","source":"Crossref","is-referenced-by-count":7,"title":["Unified Understanding of Environment, Task, and Human for Human-Robot Interaction in Real-World Environments"],"prefix":"10.1109","author":[{"given":"Yuga","family":"Yano","sequence":"first","affiliation":[{"name":"Kyushu Institute of Technology,Fukuoka,Japan"}]},{"given":"Akinobu","family":"Mizutani","sequence":"additional","affiliation":[{"name":"Kyushu Institute of Technology,Fukuoka,Japan"}]},{"given":"Yukiya","family":"Fukuda","sequence":"additional","affiliation":[{"name":"Kyushu Institute of Technology,Fukuoka,Japan"}]},{"given":"Daiju","family":"Kanaoka","sequence":"additional","affiliation":[{"name":"Kyushu Institute of Technology,Fukuoka,Japan"}]},{"given":"Tomohiro","family":"Ono","sequence":"additional","affiliation":[{"name":"Kyushu Institute of Technology,Fukuoka,Japan"}]},{"given":"Hakaru","family":"Tamukoh","sequence":"additional","affiliation":[{"name":"Kyushu Institute of Technology,Fukuoka,Japan"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-55015-7_21"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1080\/01691864.2022.2115315"},{"key":"ref3","article-title":"Real-world robotics a new principle for robotics that flexibly adapts to open environments","volume-title":"Center for Research and Development Strategy"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794456"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA40945.2020.9196885"},{"key":"ref6","article-title":"Do As I Can, Not As I Say: Grounding language in robotic affordances","author":"Ahn","year":"2022"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/iros40897.2019.8968455"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.4324\/9780203978948-13"},{"key":"ref9","article-title":"RoboCup@Home"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1186\/s40648-019-0132-3"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01264"},{"key":"ref12","article-title":"Mmdetection: Open mmlab detection toolbox and benchmark","author":"Chen","year":"2019"},{"key":"ref13","article-title":"Openmmlab pose estimation toolbox and benchmark","year":"2020"},{"key":"ref14","article-title":"OpenAI GPT-4"},{"key":"ref15","article-title":"lang-segment-anything"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.257"},{"key":"ref17","article-title":"Silero VAD: pre-trained enterprise-grade voice activity detector (VAD), number detector and language classifier","author":"Team"},{"key":"ref18","first-page":"28492","article-title":"Robust speech recognition via large-scale weak supervision","volume-title":"Proceedings of the 40th International Conference on Machine Learning (ICML)","author":"Radford"},{"key":"ref19","article-title":"Pose Anything: A Graph-Based Approach for Category-Agnostic Pose Estimation","author":"Hirschorn"},{"key":"ref20","article-title":"GPT4All: Training an assistant-style chatbot with large scale data distillation from GPT-3.5-Turbo","author":"Anand"}],"event":{"name":"2024 33rd IEEE International Conference on Robot and Human Interactive Communication (ROMAN)","location":"Pasadena, CA, USA","start":{"date-parts":[[2024,8,26]]},"end":{"date-parts":[[2024,8,30]]}},"container-title":["2024 33rd IEEE International Conference on Robot and Human Interactive Communication (ROMAN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10731142\/10731137\/10731235.pdf?arnumber=10731235","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T04:14:01Z","timestamp":1732680841000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10731235\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,26]]},"references-count":20,"URL":"https:\/\/doi.org\/10.1109\/ro-man60168.2024.10731235","relation":{},"subject":[],"published":{"date-parts":[[2024,8,26]]}}}