{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,26]],"date-time":"2025-10-26T12:08:04Z","timestamp":1761480484617,"version":"build-2065373602"},"publisher-location":"Singapore","reference-count":15,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819520978"},{"type":"electronic","value":"9789819520985"}],"license":[{"start":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T00:00:00Z","timestamp":1761523200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T00:00:00Z","timestamp":1761523200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-2098-5_5","type":"book-chapter","created":{"date-parts":[[2025,10,26]],"date-time":"2025-10-26T12:00:44Z","timestamp":1761480044000},"page":"54-66","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Soft-Skin Facial Robot Capable of\u00a0Real-Time Emotion-Driven Actuation Through Visual Perception"],"prefix":"10.1007","author":[{"given":"Xuanhe","family":"Fan","sequence":"first","affiliation":[]},{"given":"Huijuan","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Shuangjiang","family":"He","sequence":"additional","affiliation":[]},{"given":"Li","family":"Li","sequence":"additional","affiliation":[]},{"given":"Li","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"5_CR1","doi-asserted-by":"publisher","unstructured":"Chevalier, P., Kompatsiari, K., Ciardo, F., Wykowska, A.: Examining joint attention with the use of humanoid robots-a new approach to study fundamental mechanisms of social cognition. Psychon. Bull. Rev. 27(2), 217\u2013236 (2019). https:\/\/doi.org\/10.3758\/s13423-019-01689-4","DOI":"10.3758\/s13423-019-01689-4"},{"key":"5_CR2","doi-asserted-by":"publisher","unstructured":"Jo, S., Hong, S.: The development of human-robot interaction design for optimal emotional expression in social robots used by older people: design of robot facial expressions and gestures. IEEE Access 13, 21367\u201321381 (2025). https:\/\/doi.org\/10.1109\/ACCESS.2025.3534845","DOI":"10.1109\/ACCESS.2025.3534845"},{"issue":"2","key":"5_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2546276","volume":"33","author":"AH Bermano","year":"2014","unstructured":"Bermano, A.H., et al.: Facial performance enhancement using dynamic shape space analysis. ACMTrans. Graph. (TOG) 33(2), 1\u201312 (2014). https:\/\/doi.org\/10.1145\/2546276","journal-title":"ACMTrans. Graph. (TOG)"},{"key":"5_CR4","doi-asserted-by":"publisher","unstructured":"Field, M., Stirling, D., Pan, Z., Naghdy, F.: Learning trajectories for robot programing by demonstration using a coordinated mixture of factor analyzers. IEEE Trans. Cybern. 46(3), 706\u2013717 (2016). https:\/\/doi.org\/10.1109\/TCYB.2015.2414277","DOI":"10.1109\/TCYB.2015.2414277"},{"key":"5_CR5","doi-asserted-by":"publisher","unstructured":"Liu, X., Chen, Y., Li, J., Cangelosi, A.: Real-time robotic mirrored behavior of facial expressions and head motions based on lightweight networks. IEEE Internet Things J. 10(2), 1401\u20131413 (2023). https:\/\/doi.org\/10.1109\/JIOT.2022.3205123","DOI":"10.1109\/JIOT.2022.3205123"},{"key":"5_CR6","doi-asserted-by":"publisher","unstructured":"Hashimoto, T., Hiramatsu, S., Kobayashi, H.: Development of face robot for emotional communication between human and robot. In: 2006 International Conference on Mechatronics and Automation, pp. 25\u201330. IEEE, Luoyang, China (2006). https:\/\/doi.org\/10.1109\/ICMA.2006.257429","DOI":"10.1109\/ICMA.2006.257429"},{"key":"5_CR7","doi-asserted-by":"publisher","unstructured":"Hu, Y., et al.: Human-robot facial coexpression. Sci. Robot. 9, eadi4724 (2024). https:\/\/doi.org\/10.1126\/scirobotics.adi4724","DOI":"10.1126\/scirobotics.adi4724"},{"issue":"3","key":"5_CR8","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1017\/S0263574714001787","volume":"34","author":"CY Lin","year":"2016","unstructured":"Lin, C.Y., Huang, C.C., Cheng, L.C.: An expressional simplified mechanism in anthropomorphic face robot design. Robotica 34(3), 652\u2013670 (2016). https:\/\/doi.org\/10.1017\/S0263574714001787","journal-title":"Robotica"},{"issue":"1","key":"5_CR9","doi-asserted-by":"publisher","first-page":"29","DOI":"10.5772\/62181","volume":"13","author":"WT Asheber","year":"2016","unstructured":"Asheber, W.T., Lin, C.Y., Yen, S.H.: Humanoid head face mechanism with expandable facial expressions. Int. J. Adv. Rob. Syst. 13(1), 29 (2016). https:\/\/doi.org\/10.5772\/62181","journal-title":"Int. J. Adv. Rob. Syst."},{"key":"5_CR10","doi-asserted-by":"publisher","unstructured":"Randhavane, T., Bera, A., Kapsaskis, K., Sheth, R., Gray, K., Manocha, D.: EVA: generating emotional behavior of virtual agents using expressive features of gait and gaze. In: ACM Symposium on Applied Perception 2019 (SAP 2019), pp. 1\u201310. New York, USA (2019). https:\/\/doi.org\/10.1145\/3343036.3343129","DOI":"10.1145\/3343036.3343129"},{"key":"5_CR11","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.125492","volume":"262","author":"N Liu","year":"2025","unstructured":"Liu, N., et al.: A lightweight network-based sign language robot with facial mirroring and speech system. Expert Syst. Appl. 262, 125492 (2025). https:\/\/doi.org\/10.1016\/j.eswa.2024.125492","journal-title":"Expert Syst. Appl."},{"key":"5_CR12","doi-asserted-by":"publisher","unstructured":"Chen, B., Hu, Y., Li, L., Cummings, S., Lipson, H.: Smile like you mean it: driving animatronic robotic face with learned models. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 2739\u20132746. IEEE, Xi\u2019an, China (2021). https:\/\/doi.org\/10.1109\/ICRA48506.2021.9560797","DOI":"10.1109\/ICRA48506.2021.9560797"},{"key":"5_CR13","doi-asserted-by":"publisher","unstructured":"Ekman, P., Rosenberg, E.L.: What the face reveals: basic and applied studies of spontaneous expression using the facial action coding system (FACS). Oxford University Press, New York (2005). https:\/\/doi.org\/10.1093\/acprof:oso\/9780195179644.001.0001","DOI":"10.1093\/acprof:oso\/9780195179644.001.0001"},{"key":"5_CR14","doi-asserted-by":"publisher","unstructured":"Lugaresi, C., et al.: MediaPipe: a framework for building perception pipelines. arXiv preprint arXiv:1906.08172 (2019). https:\/\/doi.org\/10.48550\/arXiv.1906.08172","DOI":"10.48550\/arXiv.1906.08172"},{"key":"5_CR15","doi-asserted-by":"publisher","unstructured":"Park, J.W., Lee, H.S., Chung, M.J.: Generation of realistic robot facial expressions for human robot interaction. J. Intell. Robot. Syst. 2, 443\u2013462 (2014). https:\/\/doi.org\/10.1007\/s10846-014-0066-1","DOI":"10.1007\/s10846-014-0066-1"}],"container-title":["Lecture Notes in Computer Science","Intelligent Robotics and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-2098-5_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,26]],"date-time":"2025-10-26T12:00:44Z","timestamp":1761480044000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-2098-5_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"ISBN":["9789819520978","9789819520985"],"references-count":15,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-2098-5_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"27 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIRA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Robotics and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Okayama","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 August 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 August 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icira2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.icira2025.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}