{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T04:09:21Z","timestamp":1748664561490,"version":"3.41.0"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031929762","type":"print"},{"value":"9783031929779","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-92977-9_6","type":"book-chapter","created":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T16:24:27Z","timestamp":1748622267000},"page":"89-100","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Facial Expression Generation Based on\u00a0Custom Emotional Language and\u00a0Foundation Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2877-5380","authenticated-orcid":false,"given":"Leon","family":"Koren","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5160-9172","authenticated-orcid":false,"given":"Tomislav","family":"Stipancic","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0003-0736-1677","authenticated-orcid":false,"given":"Matija","family":"Zidaric","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5173-9396","authenticated-orcid":false,"given":"Luka","family":"Orsag","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,25]]},"reference":[{"key":"6_CR1","unstructured":"Bommasani, R., Hudson, D., Adeli, E., et\u00a0al.: On the opportunities and risks of foundation models (2021)"},{"key":"6_CR2","unstructured":"Brown, T., et\u00a0al.: Language models are few-shot learners. In: Advances in Neural Information Processing Systems, vol.\u00a033, pp. 1877\u20131901 (2020)"},{"key":"6_CR3","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman, P.: An argument for basic emotions. Cogn. Emot. 6, 169\u2013200 (1992)","journal-title":"Cogn. Emot."},{"key":"6_CR4","volume-title":"How Emotions Are Made: The Secret Life of the Brain","author":"L Feldman Barrett","year":"2017","unstructured":"Feldman Barrett, L.: How Emotions Are Made: The Secret Life of the Brain, 2nd edn. HarperCollins, New York (2017)","edition":"2"},{"key":"6_CR5","unstructured":"Feng, G., et al.: Emospeaker: One-shot fine-grained emotion-controlled talking face generation (2024)"},{"key":"6_CR6","unstructured":"Feng, X., Zhang, J., Chen, C., et\u00a0al.: Adaptive empathic voice interaction. Artificial Intelligence (2023)"},{"key":"6_CR7","doi-asserted-by":"publisher","unstructured":"Kaur, M., Kumar, M.: Facial emotion recognition: a comprehensive review. Expert Syst. 41, e13670 (2024). https:\/\/doi.org\/10.1111\/exsy.13670","DOI":"10.1111\/exsy.13670"},{"key":"6_CR8","doi-asserted-by":"publisher","unstructured":"Koren, L., Stipancic, T., Ricko, A., Orsag, L.: Multimodal emotion analysis based on visual, acoustic and linguistic features. In: Meiselwitz, G. (eds) Social Computing and Social Media: Design, User Experience and Impact. HCII 2022. LNCS, vol. 13315, pp. 318\u2013331. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-05061-9_23","DOI":"10.1007\/978-3-031-05061-9_23"},{"key":"6_CR9","doi-asserted-by":"publisher","unstructured":"Koren, L., Stipancic, T., Ricko, A., Orsag, L.: Person localization model based on a fusion of acoustic and visual inputs. Electronics 11, 440 (2022). https:\/\/doi.org\/10.3390\/ELECTRONICS11030440","DOI":"10.3390\/ELECTRONICS11030440"},{"key":"6_CR10","doi-asserted-by":"publisher","unstructured":"Koren, L., Stipancic, T., Ricko, A., Orsag, L.: Multimodal emotion analysis based on visual, acoustic and linguistic features. In: Meiselwitz, G. (eds) Social Computing and Social Media: Design, User Experience and Impact. HCII 2022. LNCS, vol. 13315, pp. 318\u2013331. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-05061-9_23","DOI":"10.1007\/978-3-031-05061-9_23"},{"key":"6_CR11","doi-asserted-by":"publisher","unstructured":"Koren, L., Stipan\u010di\u0107, T., Ri\u010dko, A., Beni\u0107, J.: Context-driven method in realization of optimized human-robot interaction. Tehni\u010dki Glasnik 16, 320\u2013327 (2022). https:\/\/doi.org\/10.31803\/tg-20220504100707","DOI":"10.31803\/tg-20220504100707"},{"key":"6_CR12","doi-asserted-by":"publisher","unstructured":"Orsag, L., Stipan\u010di\u0107, T., Koren, L.: Towards a safe human-robot collaboration using information on human worker activity. Sensors 23, 1283 (2023). https:\/\/doi.org\/10.3390\/s23031283","DOI":"10.3390\/s23031283"},{"key":"6_CR13","doi-asserted-by":"publisher","unstructured":"Orsag, L., Stipan\u010di\u0107, T., Koren, L., Posavec, K.: Human intention recognition for safe robot action planning using head pose. In: Meiselwitz, G. (eds.) Social Computing and Social Media: Design, User Experience and Impact. HCII 2022. LNCS, vol. 13315, pp. 313\u2013327. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-05061-9_23","DOI":"10.1007\/978-3-031-05061-9_23"},{"key":"6_CR14","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/1140.001.0001","volume-title":"Affective Computing","author":"R Picard","year":"1997","unstructured":"Picard, R.: Affective Computing. MIT Press, Cambridge, MA (1997)"},{"key":"6_CR15","doi-asserted-by":"publisher","first-page":"1113","DOI":"10.1109\/TPAMI.2014.2366127","volume":"37","author":"E Sariyanidi","year":"2015","unstructured":"Sariyanidi, E., Gunes, H., Cavallaro, A.: Automatic analysis of facial affect: a survey of registration, representation, and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37, 1113\u20131133 (2015). https:\/\/doi.org\/10.1109\/TPAMI.2014.2366127","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"6_CR16","doi-asserted-by":"publisher","unstructured":"Stipancic, T., Koren, L., Korade, D., Rosenberg, D.: Plea: a social robot with teaching and interacting capabilities. J. Pacific Rim Psychol. 15, 183440921137019 (2021). https:\/\/doi.org\/10.1177\/18344909211037019","DOI":"10.1177\/18344909211037019"},{"key":"6_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"279","DOI":"10.1007\/978-3-540-85891-1_31","volume-title":"Visual Information Systems. Web-Based Visual Information Search and Management","author":"P Biswas","year":"2008","unstructured":"Biswas, P., Sezgin, T.M., Robinson, P.: Perception model for people with visual impairments. In: Sebillo, M., Vitiello, G., Schaefer, G. (eds.) VISUAL 2008. LNCS, vol. 5188, pp. 279\u2013290. Springer, Heidelberg (2008). https:\/\/doi.org\/10.1007\/978-3-540-85891-1_31"},{"key":"6_CR18","unstructured":"Tsai, Y.H., Liang, P., Zadeh, A., Morency, L.P., Salakhutdinov, R.: Learning factorized multimodal representations. In: Proceedings of ICLR (2019)"},{"key":"6_CR19","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/TPAMI.2008.52","volume":"31","author":"Z Zeng","year":"2009","unstructured":"Zeng, Z., Pantic, M., Roisman, G., Huang, T.: A survey of affect recognition methods: audio, visual, and spontaneous expressions. IEEE Trans. Pattern Anal. Mach. Intell. 31, 39\u201358 (2009). https:\/\/doi.org\/10.1109\/TPAMI.2008.52","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Cao, Y., Gong, S., Patras, I.: Enhancing zero-shot facial expression recognition by LLM knowledge transfer (2024)","DOI":"10.1109\/WACV61041.2025.00089"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Zhong, Y., Wei, H., Yang, P., Wang, Z.: Expclip: bridging text and facial expressions via semantic alignment (2023)","DOI":"10.1609\/aaai.v38i7.28594"}],"container-title":["Lecture Notes in Computer Science","Distributed, Ambient and Pervasive Interactions"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-92977-9_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T16:24:31Z","timestamp":1748622271000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-92977-9_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031929762","9783031929779"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-92977-9_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"25 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"HCII","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Gothenburg","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Sweden","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 June 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 June 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"hcii2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/2025.hci.international\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}