{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T11:01:53Z","timestamp":1772276513303,"version":"3.50.1"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2026,1,11]],"date-time":"2026-01-11T00:00:00Z","timestamp":1768089600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,1,11]],"date-time":"2026-01-11T00:00:00Z","timestamp":1768089600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026,1,11]]},"DOI":"10.1109\/sii64115.2026.11404643","type":"proceedings-article","created":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T20:47:13Z","timestamp":1772225233000},"page":"1807-1812","source":"Crossref","is-referenced-by-count":0,"title":["Text-to-Motion Generation for Diverse Human Body-Motion Simulation"],"prefix":"10.1109","author":[{"given":"Jingze","family":"Gong","sequence":"first","affiliation":[{"name":"The University of Tokyo,Precision Engineering, Graduate School of Engineering,Japan"}]},{"given":"Yusheng","family":"Wang","sequence":"additional","affiliation":[{"name":"The University of Tokyo,Research Into Artifacts, Center for Engineering, Graduate School of Engineering,Japan"}]},{"given":"Jun","family":"Ota","sequence":"additional","affiliation":[{"name":"The University of Tokyo,Research Into Artifacts, Center for Engineering, Graduate School of Engineering,Japan"}]}],"member":"263","reference":[{"key":"ref1","first-page":"414","article-title":"Teach: Temporal action composition for 3d humans","volume-title":"2022 International Conference on 3D Vision (3DV)","author":"Athanasiou"},{"key":"ref2","first-page":"9760","article-title":"Mofusion: A framework for denoising-diffusion-based motion synthesis","volume-title":"2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Dabral"},{"key":"ref3","article-title":"Motion-x: A large-scale 3d expressive whole-body human motion dataset","author":"de Lin","year":"2023"},{"key":"ref4","first-page":"104450","article-title":"Motion planning in dynamic environments using context-aware human trajectory prediction","volume-title":"Robotics and Autonomous Systems","volume":"166","author":"Finean","year":"2023"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00509"},{"key":"ref6","doi-asserted-by":"crossref","DOI":"10.1145\/3394171.3413635","article-title":"Action2motion: Conditioned generation of 3d human motions","volume-title":"Proceedings of the 28th ACM International Conference on Multimedia","author":"Guo"},{"key":"ref7","first-page":"4316","article-title":"Human poseitioning system (hps): 3d human pose estimation and self-localization in large scenes from body-mounted sensors","volume-title":"2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Guzov"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01118"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00237"},{"key":"ref10","article-title":"Motiongpt: Human motion as a foreign language","volume":"36","author":"Jiang","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref11","article-title":"Act as you wish: Fine-grained control of motion diffusion model with hierarchical semantic graphs","volume":"36","author":"Jin","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref12","first-page":"19 750","article-title":"Egohumans: An egocentric 3d multi-human benchmark","volume-title":"2023 IEEE\/CVF International Conference on Computer Vision (ICCV)","author":"Khirodkar"},{"key":"ref13","article-title":"Parahome: Parameterizing everyday home activities towards 3d generative modeling of human-object interactions","volume-title":"Computer Vision and Pattern Recognition","author":"Kim","year":"2024"},{"key":"ref14","article-title":"Multiact: Long-term 3d human motion generation from multiple action labels","author":"Lee","year":"2022"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/2816795.2818013"},{"key":"ref16","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-031-72691-0_25","article-title":"Nymeria: A massive collection of multimodal egocentric daily motion in the wild","volume-title":"European Conference on Computer Vision","author":"Ma"},{"key":"ref17","first-page":"10 965","article-title":"Action-conditioned 3d human motion synthesis with transformer vae","volume-title":"2021 IEEE\/CVF International Conference on Computer Vision (ICCV)","author":"Petrovich"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1089\/big.2016.0028"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00078"},{"key":"ref20","article-title":"Fg-mdm: Towards zero-shot human motion generation via chatgpt-refined descriptions","author":"Shi","year":"2023"},{"key":"ref21","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-031-20047-2_21","article-title":"Motionclip: Exposing human motion generation to clip space","volume-title":"European Conference on Computer Vision","author":"Tevet"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20047-2_21"},{"key":"ref23","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2023.emnlp-industry.13","article-title":"Samp: A model inference toolkit of post-training quantization for text processing via self-adaptive mixed-precision","volume-title":"Conference on Empirical Methods in Natural Language Processing","author":"Tian"},{"key":"ref24","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023"},{"key":"ref25","article-title":"Neural discrete representation learning","volume-title":"Neural Information Processing Systems","author":"van den Oord","year":"2017"},{"key":"ref26","article-title":"Motion-agent: A conversational framework for human motion generation with llms","author":"Wu","year":"2024"},{"key":"ref27","first-page":"2250","article-title":"Reli11d: A comprehensive multimodal human motion dataset and method","volume-title":"2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Yan"},{"key":"ref28","article-title":"Divatrack: Diverse bodies and motions from acceleration-enhanced three-point trackers","volume-title":"Computer Graphics Forum","volume":"43","author":"Yang","year":"2024"},{"key":"ref29","first-page":"14 730","article-title":"Generating human motion from textual descriptions with discrete representations","volume-title":"2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Zhang"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2024.3355414\/mm1"},{"key":"ref31","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-031-20068-7_11","article-title":"Egobody: Human body shape and motion of interacting people from head-mounted devices","volume-title":"European Conference on Computer Vision","author":"Zhang"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01115"}],"event":{"name":"2026 IEEE\/SICE International Symposium on System Integration (SII)","location":"Cancun, Mexico","start":{"date-parts":[[2026,1,11]]},"end":{"date-parts":[[2026,1,14]]}},"container-title":["2026 IEEE\/SICE International Symposium on System Integration (SII)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11404435\/11404394\/11404643.pdf?arnumber=11404643","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T06:33:57Z","timestamp":1772260437000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11404643\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,11]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/sii64115.2026.11404643","relation":{},"subject":[],"published":{"date-parts":[[2026,1,11]]}}}