{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,16]],"date-time":"2026-02-16T16:19:51Z","timestamp":1771258791209,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":22,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819561223","type":"print"},{"value":"9789819561230","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-6123-0_18","type":"book-chapter","created":{"date-parts":[[2026,2,16]],"date-time":"2026-02-16T15:43:55Z","timestamp":1771256635000},"page":"183-194","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Online Emotion-Driven Generation of\u00a0Multiple Appropriate Facial Reactions"],"prefix":"10.1007","author":[{"given":"Jiajian","family":"Huang","sequence":"first","affiliation":[]},{"given":"Siyang","family":"Song","sequence":"additional","affiliation":[]},{"given":"Xiangyu","family":"Kong","sequence":"additional","affiliation":[]},{"given":"Weicheng","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Linlin","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Zitong","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,17]]},"reference":[{"key":"18_CR1","unstructured":"Song, S., Spitale, M., Luo, Y., et\u00a0al.: Multiple appropriate facial reaction generation in dyadic interaction settings: what, why and how? arXiv (2023)"},{"key":"18_CR2","doi-asserted-by":"crossref","unstructured":"Wang, I., Ruiz, J.: Examining the use of nonverbal communication in virtual agents. IJHCI (2021)","DOI":"10.1080\/10447318.2021.1898851"},{"key":"18_CR3","doi-asserted-by":"crossref","unstructured":"Aburumman, N., Gillies, M., Ward, J., Hamilton, A.: Nonverbal communication in virtual reality: nodding as a social signal in virtual interactions. IJHCS (2022)","DOI":"10.1016\/j.ijhcs.2022.102819"},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Hoque, X., Mann, A., Sharma, G., Dhall, A.: Beamer: behavioral encoder to generate multiple appropriate facial reactions. In: ACM MM (2023)","DOI":"10.1145\/3581783.3612860"},{"key":"18_CR5","doi-asserted-by":"crossref","unstructured":"Luo, C., Song, S., Xie, W., et\u00a0al.: ReactFace: online multiple appropriate facial reaction generation in dyadic interactions. IEEE TVCG (2024)","DOI":"10.1109\/TVCG.2024.3490613"},{"key":"18_CR6","doi-asserted-by":"crossref","unstructured":"Liang, C., Wang, J., Zhang, H., et\u00a0al.: UniFaRN: unified transformer for facial reaction generation. In: ACM MM (2023)","DOI":"10.1145\/3581783.3612854"},{"key":"18_CR7","doi-asserted-by":"crossref","unstructured":"Yu, J., Zhao, J., Xie, G., et\u00a0al.: Leveraging the latent diffusion models for offline facial multiple appropriate reactions generation. In: ACM MM (2023)","DOI":"10.1145\/3581783.3612865"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Nguyen, M., Yang, H., Ho, N., Kim, S., et\u00a0al.: Vector quantized diffusion models for multiple appropriate reactions generation. In: FG (2024)","DOI":"10.1109\/FG59268.2024.10581978"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Liu, Z., Liang, C., Wang, J., et\u00a0al.: One-to-many appropriate reaction mapping modeling with discrete latent variable. In: FG (2024)","DOI":"10.1109\/FG59268.2024.10582013"},{"key":"18_CR10","doi-asserted-by":"crossref","unstructured":"Nguyen, M., Yang, H., Kim, S., Shin, J., Kim, S.: Latent behavior diffusion for sequential reaction generation in dyadic setting. In: ICPR (2024)","DOI":"10.1007\/978-3-031-78389-0_16"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Nguyen, D., Paudel, P., Kim, S., et\u00a0al.: Multiple facial reaction generation using gaussian mixture of models and multimodal bottleneck transformer. In: FG (2024)","DOI":"10.1109\/FG59268.2024.10581901"},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"Li, J., Wang, S., Wang, X., et\u00a0al.: ReactDiff: latent diffusion for facial reaction generation. NN (2025)","DOI":"10.1016\/j.neunet.2025.107596"},{"key":"18_CR13","doi-asserted-by":"crossref","unstructured":"Lv, Q., Liu, X., Li, J., et\u00a0al.: Hierarchical multimodal decoupling-fusion framework for offline multiple appropriate facial reaction generation. In: ICASSP (2025)","DOI":"10.1109\/ICASSP49660.2025.10887743"},{"key":"18_CR14","doi-asserted-by":"crossref","unstructured":"Ma, Z., Zheng, Z., Ye, J., et\u00a0al.: Emotion2vec: self-supervised pre-training for speech emotion representation. In: ACL 2024 Findings (2024)","DOI":"10.18653\/v1\/2024.findings-acl.931"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Savchenko, A., Savchenko, L., Makarov, I.: Classifying emotions and engagement in online learning based on a single facial expression recognition neural network. IEEE TAFFC (2022)","DOI":"10.1109\/TAFFC.2022.3188390"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Song, S., Spitale, M., Luo, C., et\u00a0al.: REACT 2024: the second multiple appropriate facial reaction generation challenge. In: FG (2024)","DOI":"10.1109\/FG59268.2024.10581935"},{"key":"18_CR17","unstructured":"Xu, T., Spitale, M., Tang, H., et\u00a0al.: Reversible graph neural network-based reaction distribution learning for multiple appropriate facial reactions generation. IEEE TAFFC (2024)"},{"key":"18_CR18","doi-asserted-by":"crossref","unstructured":"Song, S., Micol, S., Kong, X., et\u00a0al.: REACT 2025: the third multiple appropriate facial reaction generation challenge. In: ACM MM (2025)","DOI":"10.1145\/3746027.3762244"},{"key":"18_CR19","doi-asserted-by":"crossref","unstructured":"Cafaro, A., Wagner, J., Baur, T., et\u00a0al.: The noxi database: multimodal recordings of mediated novice-expert interactions. In: ACM MM (2017)","DOI":"10.1145\/3136755.3136780"},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Ringeval, F., Sonderegger, A., Sauer, J., Lalanne, D.: Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions. In: FG (2013)","DOI":"10.1109\/FG.2013.6553805"},{"key":"18_CR21","doi-asserted-by":"crossref","unstructured":"Song, S., Spitale, M., Luo, C., et\u00a0al.: REACT2023: the first multiple appropriate facial reaction generation challenge. In: ACM MM (2023)","DOI":"10.1145\/3581783.3612832"},{"key":"18_CR22","doi-asserted-by":"crossref","unstructured":"Xie, X., Cui, Y., Tan, T., et\u00a0al.: FusionMamba: dynamic feature enhancement for multimodal image fusion with mamba. Vis. Intell. (2024)","DOI":"10.1007\/s44267-024-00072-9"}],"container-title":["Lecture Notes in Computer Science","Biometric Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-6123-0_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,16]],"date-time":"2026-02-16T15:44:01Z","timestamp":1771256641000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-6123-0_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819561223","9789819561230"],"references-count":22,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-6123-0_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"17 February 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CCBR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Biometric Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Nanchang","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 November 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccbr2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ccbr99.cn\/index.html","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}