{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T03:47:21Z","timestamp":1770695241333,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":32,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819569564","type":"print"},{"value":"9789819569571","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-6957-1_35","type":"book-chapter","created":{"date-parts":[[2026,2,9]],"date-time":"2026-02-09T10:44:29Z","timestamp":1770633869000},"page":"490-503","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MotionSlim: A Lightweight T2M Generation Framework Based on LLM"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-8810-8859","authenticated-orcid":false,"given":"Congrui","family":"Yu","sequence":"first","affiliation":[]},{"given":"Bo","family":"Fan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2017-2518","authenticated-orcid":false,"given":"Na","family":"Lyu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,10]]},"reference":[{"key":"35_CR1","unstructured":"Touvron, H., et al.: LLaMA: Open and efficient foundation language models (2023). arXiv preprint arXiv:2302.13971"},{"key":"35_CR2","unstructured":"Bi, X., et al.: DeepSeek LLM: scaling open-source language models with longtermism (2024). arXiv preprint arXiv:2401.02954"},{"key":"35_CR3","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: MotionGPT: finetuned LLMs are general-purpose motion generators. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 7, pp. 7368\u20137376 (2024)","DOI":"10.1609\/aaai.v38i7.28567"},{"key":"35_CR4","doi-asserted-by":"crossref","unstructured":"Ahmadian, A., et al.: Back to basics: revisiting reinforce style optimization for learning from human feedback in LLMs (2024). arXiv preprint arXiv:2402.14740","DOI":"10.18653\/v1\/2024.acl-long.662"},{"key":"35_CR5","doi-asserted-by":"crossref","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, vol. 1 (long and short papers), pp. 4171\u20134186 (2019)","DOI":"10.18653\/v1\/N19-1423"},{"issue":"140","key":"35_CR6","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"35_CR7","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)"},{"issue":"2","key":"35_CR8","first-page":"3","volume":"1","author":"EJ Hu","year":"2022","unstructured":"Hu, E.J., et al.: LoRA: low-rank adaptation of large language models. ICLR 1(2), 3 (2022)","journal-title":"ICLR"},{"key":"35_CR9","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang, L., et al.: Training language models to follow instructions with human feedback. Adv. Neural. Inf. Process. Syst. 35, 27730\u201327744 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"35_CR10","unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms (2017). arXiv preprint arXiv:1707.06347"},{"key":"35_CR11","first-page":"53728","volume":"36","author":"R Rafailov","year":"2023","unstructured":"Rafailov, R., Sharma, A., Mitchell, E., Manning, C.D., Ermon, S., Finn, C.: Direct preference optimization: your language model is secretly a reward model. Adv. Neural. Inf. Process. Syst. 36, 53728\u201353741 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"35_CR12","unstructured":"Yu, Q., et al.: DAPO: an open-source LLM reinforcement learning system at scale (2025). arXiv preprint arXiv:2503.14476"},{"key":"35_CR13","unstructured":"Shao, Z., et al.: DeepSeekMath: pushing the limits of mathematical reasoning in open language models (2024). arXiv preprint arXiv:2402.03300"},{"key":"35_CR14","doi-asserted-by":"crossref","unstructured":"Ahn, H., Ha, T., Choi, Y., Yoo, H., Oh, S.: Text2Action: generative adversarial synthesis from language to action. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 5915\u20135920. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8460608"},{"key":"35_CR15","doi-asserted-by":"crossref","unstructured":"Ghosh, A., Cheema, N., Oguz, C., Theobalt, C., Slusallek, P.: Synthesis of compositional animations from textual descriptions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1396\u20131406 (2021)","DOI":"10.1109\/ICCV48922.2021.00143"},{"key":"35_CR16","doi-asserted-by":"crossref","unstructured":"Qi, X., et al.: Weakly-supervised emotion transition learning for diverse 3D co-speech gesture generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10424\u201310434 (2024)","DOI":"10.1109\/CVPR52733.2024.00992"},{"key":"35_CR17","unstructured":"Tevet, G., Raab, S., Gordon, B., Shafir, Y., Cohen-Or, D., Bermano, A. H.: Human motion diffusion model (2022). arXiv preprint arXiv:2209.14916"},{"key":"35_CR18","unstructured":"Shafir, Y., Tevet, G., Kapon, R., Bermano, A.H.: Human motion diffusion as a generative prior (2023). arXiv preprint arXiv:2303.01418"},{"key":"35_CR19","doi-asserted-by":"publisher","unstructured":"Zhou, W., et al.: EMDM: Efficient motion diffusion model for fast and high-quality motion generation. In: Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G. (eds) European Conference on Computer Vision, pp. 18\u201338. Springer, Cham (2024). https:\/\/doi.org\/10.1007\/978-3-031-72627-9_2","DOI":"10.1007\/978-3-031-72627-9_2"},{"issue":"6","key":"35_CR20","doi-asserted-by":"publisher","first-page":"4115","DOI":"10.1109\/TPAMI.2024.3355414","volume":"46","author":"M Zhang","year":"2024","unstructured":"Zhang, M., et al.: Motiondiffuse: Text-driven human motion generation with diffusion model. IEEE Trans. Pattern Anal. Mach. Intell. 46(6), 4115\u20134128 (2024)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"35_CR21","doi-asserted-by":"crossref","unstructured":"Guo, C., et al.: Generating diverse and natural 3D human motions from text. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5152\u20135161 (2022)","DOI":"10.1109\/CVPR52688.2022.00509"},{"key":"35_CR22","doi-asserted-by":"crossref","unstructured":"Hong, F., Zhang, M., Pan, L., Cai, Z., Yang, L., Liu, Z.: AvatarCLIP: zero-shot text-driven generation and animation of 3D avatars (2022). arXiv preprint arXiv:2205.08535","DOI":"10.1145\/3528223.3530094"},{"key":"35_CR23","doi-asserted-by":"publisher","unstructured":"Petrovich, M., Black, M. J., Varol, G.: TEMOS: Generating diverse human motions from textual descriptions. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds) European Conference on Computer Vision, pp. 480\u2013497 (2022). Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-031-20047-2_28","DOI":"10.1007\/978-3-031-20047-2_28"},{"key":"35_CR24","doi-asserted-by":"publisher","unstructured":"Guo, C., Zuo, X., Wang, S., Cheng, L.: TM2T: stochastic and tokenized modeling for the reciprocal generation of 3D human motions and texts. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds) European Conference on Computer Vision, vol. 13695, pp. 580\u2013597 (2022). Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-031-19833-5_34","DOI":"10.1007\/978-3-031-19833-5_34"},{"key":"35_CR25","doi-asserted-by":"crossref","unstructured":"Athanasiou, N., Petrovich, M., Black, M. J., Varol, G.: TEACH: temporal action composition for 3D humans. In: 2022 International Conference on 3D Vision (3DV), pp. 414\u20131423. IEEE (2022)","DOI":"10.1109\/3DV57658.2022.00053"},{"key":"35_CR26","unstructured":"Wu, Q., Zhao, Y., Wang, Y., Liu, X., Tai, Y. W., Tang, C. K.: Motion-agent: a conversational framework for human motion generation with LLMs (2024). arXiv preprint arXiv:2405.17013"},{"key":"35_CR27","unstructured":"Wang, Y., et al.: MotionGPT-2: a general-purpose motion-language model for motion generation and understanding (2024). arXiv preprint arXiv:2410.21747"},{"key":"35_CR28","unstructured":"Ouyang, R., Li, H., Zhang, Z., Wang, X., Zhu, Z., Huang, G., Wang, X.: Motion-R1: chain-of-thought reasoning and reinforcement learning for human motion generation (2025). arXiv preprint arXiv:2506.10353"},{"key":"35_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, J., et al.: Generating human motion from textual descriptions with discrete representations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14730\u201314740 (2023)","DOI":"10.1109\/CVPR52729.2023.01415"},{"key":"35_CR30","doi-asserted-by":"crossref","unstructured":"Mahmood, N., Ghorbani, N., Troje, N. F., Pons-Moll, G., Black, M. J.: AMASS: archive of motion capture as surface shapes. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5442\u20135451 (2019)","DOI":"10.1109\/ICCV.2019.00554"},{"key":"35_CR31","doi-asserted-by":"crossref","unstructured":"Guo, C., et al.: Action2Motion: conditioned generation of 3d human motions. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 2021\u20132029 (2020)","DOI":"10.1145\/3394171.3413635"},{"key":"35_CR32","doi-asserted-by":"publisher","unstructured":"Jiang, B., et al.: MotionChain: Conversational motion controllers via multimodal prompts. In: Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G. (eds) European Conference on Computer Vision, pp. 54\u201374 (2024). Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-031-73347-5_4","DOI":"10.1007\/978-3-031-73347-5_4"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-6957-1_35","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,9]],"date-time":"2026-02-09T10:44:38Z","timestamp":1770633878000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-6957-1_35"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819569564","9789819569571"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-6957-1_35","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"10 February 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Prague","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Czech Republic","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2026","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2026","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31 January 2026","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"32","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2026","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/mmm2026.cz\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}