{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T08:30:22Z","timestamp":1768033822663,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":81,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,10]]},"DOI":"10.1145\/3765766.3765785","type":"proceedings-article","created":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T14:07:31Z","timestamp":1767362851000},"page":"247-256","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Impact of Personality on Generation of Co-speech Nonverbal Behaviors Represented by 3D Skeleton Pose"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-3849-1656","authenticated-orcid":false,"given":"Ryo","family":"Ishii","sequence":"first","affiliation":[{"name":"Human Informatics Laboratories, NTT, Inc., Yokosuka-shi, Kanagawa, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4351-3530","authenticated-orcid":false,"given":"Shinichiro","family":"Eitoku","sequence":"additional","affiliation":[{"name":"Human Informatics Laboratories, NTT, Inc., Yokosuka-shi, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-1914-7294","authenticated-orcid":false,"given":"Yoshihide","family":"Sato","sequence":"additional","affiliation":[{"name":"Human Informatics Laboratories, NTT, Inc., Yokosuka-shi, Japan"}]}],"member":"320","published-online":{"date-parts":[[2026,1,2]]},"reference":[{"key":"e_1_3_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.170"},{"key":"e_1_3_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01991"},{"key":"e_1_3_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58523-5_15"},{"key":"e_1_3_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340555.3353725"},{"key":"e_1_3_3_2_6_2","volume-title":"IEEE Transactions on Visualization and Computer Graphics","author":"Ahuja Karttikeya","year":"2023","unstructured":"Karttikeya Ahuja, Gustav\u00a0Eje Henter, Taras Kucherenko, and Simon Alexanderson. 2023. StyleGestures: Style-controllable expressive co-speech gesture generation. In IEEE Transactions on Visualization and Computer Graphics."},{"key":"e_1_3_3_2_7_2","doi-asserted-by":"crossref","unstructured":"Simon Alexanderson Gustav\u00a0Eje Henter Taras Kucherenko and Jonas Beskow. 2020. Style-Controllable Speech-Driven Gesture Synthesis Using Normalising Flows. Computer Graphics Forum 39 2 (2020) 487\u2013496.","DOI":"10.1111\/cgf.13946"},{"key":"e_1_3_3_2_8_2","doi-asserted-by":"crossref","unstructured":"Tenglong Ao Qingzhe Gao Yuke Lou Baoquan Chen and Libin Liu. 2022. Rhythmic Gesticulator: Rhythm-Aware Co-Speech Gesture Synthesis with Hierarchical Neural Embeddings. ACM Trans. Graph. 41 6 (2022).","DOI":"10.1145\/3550454.3555435"},{"key":"e_1_3_3_2_9_2","volume-title":"Bodily Communication","author":"Argyle Michael","year":"1988","unstructured":"Michael Argyle. 1988. Bodily Communication. Methuen Publishing."},{"key":"e_1_3_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2006-375"},{"key":"e_1_3_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1109\/VR50410.2021.00037"},{"key":"e_1_3_3_2_12_2","volume-title":"Kinesics and context","author":"Birdwhistell Ray\u00a0L.","year":"1970","unstructured":"Ray\u00a0L. Birdwhistell. 1970. Kinesics and context. Univ. Pennsylvania Press."},{"key":"e_1_3_3_2_13_2","doi-asserted-by":"crossref","unstructured":"Elif Bozkurt Y\u00fccel Yemez and Engin Erzin. 2020. Affective synthesis and animation of arm gestures from speech prosody. Speech Communication 119 (2020) 1\u201311.","DOI":"10.1016\/j.specom.2020.02.005"},{"key":"e_1_3_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-94-009-4912-62"},{"key":"e_1_3_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2006.885910"},{"key":"e_1_3_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/383259.383315"},{"key":"e_1_3_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-21996-7_17"},{"key":"e_1_3_3_2_18_2","volume-title":"Context-Awareness in Human-Robot Interaction: Approaches and Challenges, workshop at 2022 ACM\/IEEE International Conference on Human-Robot Interaction","author":"Deichler A.","year":"2022","unstructured":"A. Deichler, S. Wang, S. Alexanderson, and J. Beskow. 2022. Towards Context-Aware Human-like Pointing Gestures with RL Motion Imitation. In Context-Awareness in Human-Robot Interaction: Approaches and Challenges, workshop at 2022 ACM\/IEEE International Conference on Human-Robot Interaction."},{"key":"e_1_3_3_2_19_2","doi-asserted-by":"crossref","unstructured":"Mireille Fares Catherine Pelachaud and Nicolas Obin. 2023. Zero-shot style transfer for gesture animation driven by text and speech using adversarial disentanglement of multimodal style encoding. Frontiers in Artificial Intelligence 6 (2023).","DOI":"10.3389\/frai.2023.1142997"},{"key":"e_1_3_3_2_20_2","doi-asserted-by":"crossref","unstructured":"Ylva Ferstl Michael Neff and Rachel McDonnell. 2021. ExpressGesture: Expressive gesture generation from speech through database matching. Computer Animation and Virtual Worlds 32 3-4 (2021) e2016.","DOI":"10.1002\/cav.2016"},{"key":"e_1_3_3_2_21_2","volume-title":"Posture & Perception in the Context of the Tonic Function Model of Structural Integration: An Introduction","author":"Frank Kevin","year":"2007","unstructured":"Kevin Frank. 2007. Posture & Perception in the Context of the Tonic Function Model of Structural Integration: An Introduction. IASI Yearbook."},{"key":"e_1_3_3_2_22_2","unstructured":"Nan Gao Zeyu Zhao Zhi Zeng Shuwu Zhang Dongdong Weng and Yihua Bao. 2023. GesGPT: Speech Gesture Synthesis With Text Parsing from ChatGPT. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.13013 (2023). https:\/\/arxiv.org\/abs\/2303.13013"},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1145\/3536221.3558068"},{"key":"e_1_3_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00361"},{"key":"e_1_3_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/AFGR.2002.1004186"},{"key":"e_1_3_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1145\/3528233.3530750"},{"key":"e_1_3_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1145\/3472306.3478335"},{"key":"e_1_3_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3267851.3267878"},{"key":"e_1_3_3_2_29_2","first-page":"1784","volume-title":"Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC\u201916)","author":"Hough Julian","year":"2016","unstructured":"Julian Hough, Ye Tian, Laura de Ruiter, Simon Betz, Spyros Kousidis, David Schlangen, and Jonathan Ginzburg. 2016. DUEL: A Multi-lingual Multimodal Dialogue Corpus for Disfluency, Exclamations and Laughter. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC\u201916), Nicoletta Calzolari, Khalid Choukri, Thierry Declerck, Sara Goggi, Marko Grobelnik, Bente Maegaard, Joseph Mariani, Helene Mazo, Asuncion Moreno, Jan Odijk, and Stelios Piperidis (Eds.). European Language Resources Association (ELRA), Portoro\u017e, Slovenia, 1784\u20131788. https:\/\/aclanthology.org\/L16-1281\/"},{"key":"e_1_3_3_2_30_2","doi-asserted-by":"crossref","unstructured":"Katherine Isbister and Clifford Nass. 2000. Consistency of personality in interactive characters: Verbal cues non-verbal cues and user characteristics. J. Human-Computer Studies 53 2 (2000) 251\u2013267.","DOI":"10.1006\/ijhc.2000.0368"},{"key":"e_1_3_3_2_31_2","doi-asserted-by":"crossref","unstructured":"Carlos\u00a0T. Ishi Daichi Machiyashiki Ryusuke Mikata and Hiroshi Ishiguro. 2018. A Speech-Driven Hand Gesture Generation Method and Evaluation in Android Robots. IEEE Robotics and Automation Letters 3 4 (2018) 3757\u20133764.","DOI":"10.1109\/LRA.2018.2856281"},{"key":"e_1_3_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.1145\/3383652.3423908"},{"key":"e_1_3_3_2_33_2","doi-asserted-by":"crossref","unstructured":"Ryo Ishii Ryuichiro Higashinaka Koh Mitsuda Taichi Katayama Masahiro Mizukami Junji Tomita Hidetoshi Kawabata Emi Yamaguchi Noritake Adachi and Yushi Aono. 2021. Methods for Efficiently Constructing Text-dialogue-agent System using Existing Anime Characters. Journal of Information Processing 29 (2021) 30\u201344.","DOI":"10.2197\/ipsjjip.29.30"},{"key":"e_1_3_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1145\/3267851.3267869"},{"key":"e_1_3_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1145\/3267851.3267866"},{"key":"e_1_3_3_2_36_2","doi-asserted-by":"publisher","unstructured":"Ryo Ishii Xutong Ren Michal Muszynski and Louis\u2011Philippe Morency. 2022. Trimodal prediction of speaking and listening willingness to help improve turn-changing modeling. Frontiers in Psychology 13 (2022) 774547. 10.3389\/fpsyg.2022.774547","DOI":"10.3389\/fpsyg.2022.774547"},{"key":"e_1_3_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1145\/3383652.3423907"},{"key":"e_1_3_3_2_38_2","first-page":"131","volume-title":"International Conference on Intelligent Virtual Agents (IVA)","author":"Ishii Ryo","year":"2021","unstructured":"Ryo Ishii, Xutong Ren, Michal Muszynski, and Louis-Philippe Morency. 2021. Multimodal and multitask approach to listener\u2019s backchannel prediction: Can prediction of turn-changing and turn-management willingness improve backchannel modeling?. In International Conference on Intelligent Virtual Agents (IVA). 131\u2013138."},{"key":"e_1_3_3_2_39_2","first-page":"66","volume-title":"The \"Big Five\" Factor Taxonomy: Dimensions of Personality in the Natural Language and in Questionnaires","author":"John Oliver\u00a0P.","year":"1990","unstructured":"Oliver\u00a0P. John. 1990. The \"Big Five\" Factor Taxonomy: Dimensions of Personality in the Natural Language and in Questionnaires. Guilford Press, 66\u2013100."},{"key":"e_1_3_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1109\/HRI.2016.7451799"},{"key":"e_1_3_3_2_41_2","first-page":"133","volume-title":"Psychol Sci.","author":"KG Munhall","year":"2004","unstructured":"Munhall KG, Jeffery\u00a0A Jones, Daniel\u00a0E Callan, Takaaki Kuratate, and Eric Vatikiotis-Bateson. 2004. Visual Prosody and Speech Intelligibility: Head Movement Improves Auditory Speech Perception. In Psychol Sci. , Vol.\u00a015. 133\u2013137."},{"key":"e_1_3_3_2_42_2","unstructured":"Mark\u00a0L Knapp and Gerald\u00a0R Miller. 1994. Communicator Characteristics and Behavior. Handbook of Interpersonal Communication (1994) 103\u2013161."},{"key":"e_1_3_3_2_43_2","doi-asserted-by":"publisher","DOI":"10.1145\/3308532.3329472"},{"key":"e_1_3_3_2_44_2","doi-asserted-by":"crossref","unstructured":"Taras Kucherenko Dai Hasegawa Naoshi Kaneko Gustav\u00a0Eje Henter and Hedvig Kjellstr\u00f6m. 2021. Moving Fast and Slow: Analysis of Representations and Post-Processing in Speech-Driven Automatic Gesture Generation. International Journal of Human\u2013Computer Interaction 37 14 (2021) 1300\u20131316.","DOI":"10.1080\/10447318.2021.1883883"},{"key":"e_1_3_3_2_45_2","volume-title":"Computer Animation and Virtual Worlds","author":"Kucherenko Taras","year":"2021","unstructured":"Taras Kucherenko, Gustav\u00a0Eje Henter, Simon Alexanderson, and Hedvig Kjellstr\u00f6m. 2021. The GENEA Challenge 2020: Benchmarking gesture-generation systems on common data. In Computer Animation and Virtual Worlds."},{"key":"e_1_3_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.1145\/3382507.3418815"},{"key":"e_1_3_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.5555\/3535850.3535937"},{"key":"e_1_3_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1145\/3461615.3485408"},{"key":"e_1_3_3_2_49_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01110"},{"key":"e_1_3_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01022"},{"key":"e_1_3_3_2_51_2","doi-asserted-by":"publisher","DOI":"10.1145\/2556288.2557274"},{"key":"e_1_3_3_2_52_2","doi-asserted-by":"crossref","unstructured":"Albert Mehrabian. 1969. Significance of posture and position in the communication of attitude and status relationships. Psychological Bulletin 71 5 (1969) 359\u2013372.","DOI":"10.1037\/h0027349"},{"key":"e_1_3_3_2_53_2","doi-asserted-by":"publisher","unstructured":"Yukiko\u00a0I. Nakano Fumio Nihei Ryo Ishii and Ryuichiro Higashinaka. 2024. Selecting Iconic Gesture Forms Based on Typical Entity Images. Journal of Information Processing 32 (2024) 196\u2013205. 10.2197\/ipsjjip.32.196","DOI":"10.2197\/ipsjjip.32.196"},{"key":"e_1_3_3_2_54_2","doi-asserted-by":"publisher","DOI":"10.3115\/1613984.1614023"},{"key":"e_1_3_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-15892-6_24"},{"key":"e_1_3_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340555.3353736"},{"key":"e_1_3_3_2_57_2","volume-title":"Personality Assessment Through Movement","author":"North Marion","year":"1972","unstructured":"Marion North. 1972. Personality Assessment Through Movement. Macdonald and Evans. https:\/\/books.google.com\/books?id=pm19AAAAMAAJ"},{"key":"e_1_3_3_2_58_2","volume-title":"EUROGRAPHICS","author":"Nyatsanga Simbarashe","year":"2023","unstructured":"Simbarashe Nyatsanga, Taras Kucherenko, Chaitanya Ahuja, Gustav Henter, and Michael Neff. 2023. A Comprehensive Review of Data-Driven Co-Speech Gesture Generation. In EUROGRAPHICS , Vol.\u00a042."},{"key":"e_1_3_3_2_59_2","doi-asserted-by":"publisher","unstructured":"Catherine Pelachaud. 2009. Studies on gesture expressivity for a virtual agent. Speech Communication 51 7 (2009) 630\u2013639. 10.1016\/j.specom.2008.04.009Research Challenges in Speech Technology: A Special Issue in Honour of Rolf Carlson and Bj\u00f6rn Granstr\u00f6m.","DOI":"10.1016\/j.specom.2008.04.009"},{"key":"e_1_3_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01089"},{"key":"e_1_3_3_2_61_2","doi-asserted-by":"crossref","unstructured":"Manuel Rebol Christian G\u00fctl and Krzysztof Pietroszek. 2021. Passing a Non-verbal Turing Test: Evaluating Gesture Animations Generated from Speech. 2021 IEEE Virtual Reality and 3D User Interfaces (VR) (2021) 573\u2013581.","DOI":"10.1109\/VR50410.2021.00082"},{"key":"e_1_3_3_2_62_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-85483-8_23"},{"key":"e_1_3_3_2_63_2","doi-asserted-by":"crossref","unstructured":"RichardLippa. 1998. The Nonverbal Display and Judgment of Extraversion Masculinity Femininity and Gender Diagnosticity: A Lens Model Analysis. J. Research in Personality 32 1 (1998) 80\u2013107.","DOI":"10.1006\/jrpe.1997.2189"},{"key":"e_1_3_3_2_64_2","doi-asserted-by":"crossref","unstructured":"Ronald\u00a0E. Riggio and Howard\u00a0S. Friedman. 1986. Impression Formation: The Role of Expressive Behavior. J. Personality and Social Psychology 50 2 (1986) 421\u2013427.","DOI":"10.1037\/\/0022-3514.50.2.421"},{"key":"e_1_3_3_2_65_2","doi-asserted-by":"crossref","unstructured":"Najmeh Sadoughi and Carlos Busso. 2019. Speech-driven animation with meaningful behaviors. Speech Communication 110 (2019) 90\u2013100.","DOI":"10.1016\/j.specom.2019.04.005"},{"key":"e_1_3_3_2_66_2","unstructured":"Carolyn Saund Andrei B\u00eerl\u0103deanu and Stacy Marsella. 2021. CMCF: An Architecture for Realtime Gesture Generation by Clustering Gestures by Motion and Communicative Function. 1136\u20131144."},{"key":"e_1_3_3_2_67_2","unstructured":"Senko Maynard. 1989. Japanese conversation: Self-contextualization through structure and Interactional management. Ablex Publishing Corporation (1989)."},{"key":"e_1_3_3_2_68_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-42293-571"},{"key":"e_1_3_3_2_69_2","doi-asserted-by":"publisher","unstructured":"Harrison\u00a0Jesse Smith and Michael Neff. 2017. Understanding the Impact of Animated Gesture Performance on Personality Perceptions. ACM Trans. Graph. 36 4 Article 49 (July 2017) 12\u00a0pages. 10.1145\/3072959.3073697","DOI":"10.1145\/3072959.3073697"},{"key":"e_1_3_3_2_70_2","doi-asserted-by":"publisher","DOI":"10.5555\/3295222.3295349"},{"key":"e_1_3_3_2_71_2","doi-asserted-by":"crossref","unstructured":"Petra Wagner Zofia Malisz and Stefan Kopp. 2014. Gesture and Speech in Interaction: An Overview. Speech Communication 57 (2014) 209\u2013232.","DOI":"10.1016\/j.specom.2013.09.008"},{"key":"e_1_3_3_2_72_2","doi-asserted-by":"publisher","unstructured":"Yingying Wang Jean E.\u00a0Fox Tree Marilyn Walker and Michael Neff. 2016. Assessing the Impact of Hand Motion on Virtual Character Personality. ACM Trans. Appl. Percept. 13 2 Article 9 (March 2016) 23\u00a0pages. 10.1145\/2874357","DOI":"10.1145\/2874357"},{"key":"e_1_3_3_2_73_2","doi-asserted-by":"publisher","DOI":"10.1145\/3461615.3485407"},{"key":"e_1_3_3_2_74_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-long.273"},{"key":"e_1_3_3_2_75_2","doi-asserted-by":"crossref","unstructured":"Zijie Ye Haozhe Wu and Jia Jia. 2022. Human motion modeling with deep learning: A survey. AI Open 3 (2022) 35\u201339.","DOI":"10.1016\/j.aiopen.2021.12.002"},{"key":"e_1_3_3_2_76_2","doi-asserted-by":"crossref","unstructured":"Hani\u00a0Camille Yehia Takaaki Kuratate and Eric Vatikiotis-Bateson. 2002. Linking facial animation head motion and speech acoustics. 30 3 (2002) 555\u2013568.","DOI":"10.1006\/jpho.2002.0165"},{"key":"e_1_3_3_2_77_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8793720"},{"key":"e_1_3_3_2_78_2","doi-asserted-by":"publisher","DOI":"10.1145\/3414685.3417838"},{"key":"e_1_3_3_2_79_2","doi-asserted-by":"crossref","unstructured":"Youngwoo Yoon Pieter Wolfert Taras Kucherenko Carla Viegas Teodor Nikolov Mihail Tsakov and Gustav\u00a0Eje Henter. 2022. The GENEA Challenge 2022: A Large Evaluation of Data-Driven Co-Speech Gesture Generation(ICMI \u201922). Association for Computing Machinery 736\u2013747.","DOI":"10.1145\/3536221.3558058"},{"key":"e_1_3_3_2_80_2","doi-asserted-by":"publisher","DOI":"10.1145\/3536221.3558063"},{"key":"e_1_3_3_2_81_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00341"},{"key":"e_1_3_3_2_82_2","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/863"}],"event":{"name":"HAI '25: International Conference on Human-Agent Interaction","location":"Yokohama Japan","acronym":"HAI '25","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 13th International Conference on Human-Agent Interaction"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3765766.3765785","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,9]],"date-time":"2026-01-09T19:19:44Z","timestamp":1767986384000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3765766.3765785"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,10]]},"references-count":81,"alternative-id":["10.1145\/3765766.3765785","10.1145\/3765766"],"URL":"https:\/\/doi.org\/10.1145\/3765766.3765785","relation":{},"subject":[],"published":{"date-parts":[[2025,11,10]]},"assertion":[{"value":"2026-01-02","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}