{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T05:25:55Z","timestamp":1731993955357,"version":"3.28.0"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"S1","license":[{"start":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T00:00:00Z","timestamp":1715385600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T00:00:00Z","timestamp":1715385600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s11760-024-03202-4","type":"journal-article","created":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T10:05:21Z","timestamp":1715421921000},"page":"909-918","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Synthesizing facial expressions in dyadic human\u2013robot interaction"],"prefix":"10.1007","volume":"18","author":[{"given":"Abdallah Hussein","family":"Sham","sequence":"first","affiliation":[]},{"given":"Pia","family":"Tikka","sequence":"additional","affiliation":[]},{"given":"David","family":"Lamas","sequence":"additional","affiliation":[]},{"given":"Gholamreza","family":"Anbarjafari","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,11]]},"reference":[{"issue":"1","key":"3202_CR1","doi-asserted-by":"publisher","first-page":"5518","DOI":"10.1038\/s41598-021-84880-9","volume":"11","author":"BEF Mota","year":"2021","unstructured":"Mota, B.E.F., Rodrigues, P.O., Lacerda, K.C.D., David, I.A., Volchan, E., Campagnoli, R.R., Souza, G.G.L.: Pictures of social interaction prompt a sustained increase of the smile expression and induce sociability. Sci. Rep. 11(1), 5518 (2021)","journal-title":"Sci. Rep."},{"key":"3202_CR2","doi-asserted-by":"publisher","first-page":"362","DOI":"10.1016\/j.cortex.2020.06.010","volume":"130","author":"S Schindler","year":"2020","unstructured":"Schindler, S., Bublatzky, F.: Attention and emotion: an integrative review of emotional face processing as a function of attention. Cortex 130, 362\u2013386 (2020)","journal-title":"Cortex"},{"issue":"1","key":"3202_CR3","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1177\/1529100619850176","volume":"20","author":"A Cowen","year":"2019","unstructured":"Cowen, A., Sauter, D., Tracy, J.L., Keltner, D.: Mapping the passions: toward a high-dimensional taxonomy of emotional experience and expression. Psychol. Sci. Public Interest 20(1), 69\u201390 (2019)","journal-title":"Psychol. Sci. Public Interest"},{"issue":"1","key":"3202_CR4","doi-asserted-by":"publisher","first-page":"150","DOI":"10.1080\/02699931.2020.1815655","volume":"35","author":"AC Holland","year":"2021","unstructured":"Holland, A.C., O\u2019Connell, G., Dziobek, I.: Facial mimicry, empathy, and emotion recognition: a meta-analysis of correlations. Cogn. Emot. 35(1), 150\u2013168 (2021)","journal-title":"Cogn. Emot."},{"issue":"8","key":"3202_CR5","doi-asserted-by":"publisher","first-page":"803","DOI":"10.1093\/scan\/nsaa110","volume":"15","author":"S Volynets","year":"2020","unstructured":"Volynets, S., Smirnov, D., Saarim\u00e4ki, H., Nummenmaa, L.: Statistical pattern recognition reveals shared neural signatures for displaying and recognizing specific facial expressions. Soc. Cognitive Affect. Neurosci. 15(8), 803\u2013813 (2020)","journal-title":"Soc. Cognitive Affect. Neurosci."},{"issue":"1","key":"3202_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1177\/1529100619832930","volume":"20","author":"LF Barrett","year":"2019","unstructured":"Barrett, L.F., Adolphs, R., Marsella, S., Martinez, A.M., Pollak, S.D.: Emotional expressions reconsidered: challenges to inferring emotion from human facial movements. Psychol. Sci. Public Interest 20(1), 1\u201368 (2019)","journal-title":"Psychol. Sci. Public Interest"},{"issue":"6\u20137","key":"3202_CR7","doi-asserted-by":"publisher","first-page":"300","DOI":"10.3109\/03091902.2011.601784","volume":"35","author":"K Gouizi","year":"2011","unstructured":"Gouizi, K., Bereksi Reguig, F., Maaoui, C.: Emotion recognition from physiological signals. J. Med. Eng. Technol. 35(6\u20137), 300\u2013307 (2011)","journal-title":"J. Med. Eng. Technol."},{"key":"3202_CR8","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.entcs.2019.04.009","volume":"343","author":"M Egger","year":"2019","unstructured":"Egger, M., Ley, M., Hanke, S.: Emotion recognition from physiological signal analysis: a review. Electron. Notes Theoretical Comput. Sci. 343, 35\u201355 (2019)","journal-title":"Electron. Notes Theoretical Comput. Sci."},{"key":"3202_CR9","doi-asserted-by":"crossref","unstructured":"Seibt, J.: \u201cTowards an ontology of simulated social interaction: varieties of the \u201cas if\u201d for robots and humans,\u201d In: Sociality and normativity for robots: Philosophical inquiries into human-robot interactions. Springer, pp. 11\u201339 (2017)","DOI":"10.1007\/978-3-319-53133-5_2"},{"issue":"3","key":"3202_CR10","doi-asserted-by":"publisher","first-page":"361","DOI":"10.1007\/s11023-021-09558-7","volume":"31","author":"JK Malinowska","year":"2021","unstructured":"Malinowska, J.K.: What does it mean to empathise with a robot? Mind. Mach. 31(3), 361\u2013376 (2021)","journal-title":"Mind. Mach."},{"issue":"5","key":"3202_CR11","doi-asserted-by":"publisher","first-page":"294","DOI":"10.1089\/cyber.2020.0181","volume":"24","author":"J Fox","year":"2021","unstructured":"Fox, J., Gambino, A.: Relationship development with humanoid social robots: applying interpersonal theories to human-robot interaction. Cyberpsychol. Behav. Soc. Netw. 24(5), 294\u2013299 (2021)","journal-title":"Cyberpsychol. Behav. Soc. Netw."},{"issue":"3","key":"3202_CR12","doi-asserted-by":"publisher","first-page":"1889","DOI":"10.3390\/ijerph19031889","volume":"19","author":"S Park","year":"2022","unstructured":"Park, S., Whang, M.: Empathy in human-robot interaction: designing for social robots. Int. J. Environ. Res. Public Health 19(3), 1889 (2022)","journal-title":"Int. J. Environ. Res. Public Health"},{"key":"3202_CR13","doi-asserted-by":"publisher","first-page":"468","DOI":"10.3389\/fpsyg.2018.00468","volume":"9","author":"L Damiano","year":"2018","unstructured":"Damiano, L., Dumouchel, P.: Anthropomorphism in human-robot co-evolution. Front. Psychol. 9, 468 (2018)","journal-title":"Front. Psychol."},{"key":"3202_CR14","doi-asserted-by":"publisher","first-page":"609673","DOI":"10.3389\/frai.2020.609673","volume":"3","author":"JR Lee","year":"2021","unstructured":"Lee, J.R., Wang, L., Wong, A.: Emotionnet nano: an efficient deep convolutional neural network design for real-time facial expression recognition. Front. Artif. Intell. 3, 609673 (2021)","journal-title":"Front. Artif. Intell."},{"issue":"8","key":"3202_CR15","doi-asserted-by":"publisher","first-page":"819","DOI":"10.21037\/qims.2018.09.07","volume":"8","author":"M-Y Wang","year":"2018","unstructured":"Wang, M.-Y., Luan, P., Zhang, J., Xiang, Y.-T., Niu, H., Yuan, Z.: Concurrent mapping of brain activation from multiple subjects during social interaction by hyperscanning: a mini-review. Quant. Imaging Med. Surg. 8(8), 819 (2018)","journal-title":"Quant. Imaging Med. Surg."},{"issue":"4","key":"3202_CR16","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1016\/j.neunet.2005.03.006","volume":"18","author":"N Fragopanagos","year":"2005","unstructured":"Fragopanagos, N., Taylor, J.G.: Emotion recognition in human-computer interaction. Neural Netw. 18(4), 389\u2013405 (2005)","journal-title":"Neural Netw."},{"issue":"14","key":"3202_CR17","doi-asserted-by":"publisher","first-page":"4626","DOI":"10.3390\/s21144626","volume":"21","author":"A Toichoa Eyam","year":"2021","unstructured":"Toichoa Eyam, A., Mohammed, W.M., Martinez Lastra, J.L.: Emotion-driven analysis and control of human-robot interactions in collaborative applications. Sensors 21(14), 4626 (2021)","journal-title":"Sensors"},{"key":"3202_CR18","doi-asserted-by":"crossref","unstructured":"Spezialetti, M., Placidi, G., Rossi, S.: Emotion recognition for human-robot interaction: Recent advances and future perspectives. Front. Robotics AI, p. 145, (2020)","DOI":"10.3389\/frobt.2020.532279"},{"key":"3202_CR19","doi-asserted-by":"crossref","unstructured":"Duthoit, C.\u00a0J., Sztynda, T., Lal, S.\u00a0K., Jap, B.\u00a0T., Agbinya, J.\u00a0I.: Optical flow image analysis of facial expressions of human emotion: Forensic applications. In: Proceedings of the 1st international conference on Forensic applications and techniques in telecommunications, information, and multimedia and workshop, pp. 1\u20136 (2008)","DOI":"10.4108\/e-forensics.2008.2743"},{"key":"3202_CR20","doi-asserted-by":"crossref","unstructured":"Popa, M., Rothkrantz, L., Wiggers, P.: Products appreciation by facial expressions analysis. In: Proceedings of the 11th International Conference on Computer Systems and Technologies and Workshop for PhD Students in Computing on International Conference on Computer Systems and Technologies, pp. 293\u2013298 (2010)","DOI":"10.1145\/1839379.1839431"},{"key":"3202_CR21","doi-asserted-by":"crossref","unstructured":"Liu, X., Zhang, L., Yadegar, J.: A multi-modal emotion recognition system for persistent and non-invasive personal health monitoring. In: Proceedings of the 2nd Conference on Wireless Health, pp. 1\u20132 (2011)","DOI":"10.1145\/2077546.2077577"},{"issue":"1","key":"3202_CR22","first-page":"171","volume":"2","author":"P Metri","year":"2012","unstructured":"Metri, P., Ghorpade, J.: Facial emotion recognition using context based multimodal approach. Int. J. Emerging Sci. 2(1), 171 (2012)","journal-title":"Int. J. Emerging Sci."},{"key":"3202_CR23","doi-asserted-by":"crossref","unstructured":"Soladi\u00e9, C., Salam, H., Pelachaud, C., Stoiber, N., S\u00e9guier, R.: A multimodal fuzzy inference system using a continuous facial expression representation for emotion detection. In: Proceedings of the 14th ACM international conference on Multimodal interaction, pp. 493\u2013500 (2012)","DOI":"10.1145\/2388676.2388782"},{"key":"3202_CR24","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1007\/s00779-011-0477-y","volume":"17","author":"V Terzis","year":"2013","unstructured":"Terzis, V., Moridis, C.N., Economides, A.A.: Measuring instant emotions based on facial expressions during computer-based assessment. Pers. Ubiquit. Comput. 17, 43\u201352 (2013)","journal-title":"Pers. Ubiquit. Comput."},{"key":"3202_CR25","doi-asserted-by":"crossref","unstructured":"Zen, G., Sangineto, E., Ricci, E., Sebe, N.: Unsupervised domain adaptation for personalized facial emotion recognition. In: Proceedings of the 16th international conference on multimodal interaction, pp. 128\u2013135 (2014)","DOI":"10.1145\/2663204.2663247"},{"key":"3202_CR26","doi-asserted-by":"publisher","first-page":"26391","DOI":"10.1109\/ACCESS.2018.2831927","volume":"6","author":"J Guo","year":"2018","unstructured":"Guo, J., Lei, Z., Wan, J., Avots, E., Hajarolasvadi, N., Knyazev, B., Kuharenko, A., Junior, J.C.S.J., Bar\u00f3, X., Demirel, H., et al.: Dominant and complementary emotion recognition from still images of faces. IEEE Access 6, 26391\u201326403 (2018)","journal-title":"IEEE Access"},{"key":"3202_CR27","doi-asserted-by":"crossref","unstructured":"Slimani, K., Kas, M., El\u00a0Merabet, Y., Messoussi, R., Ruichek, Y.: Facial emotion recognition: A comparative analysis using 22 lbp variants. In: Proceedings of the 2nd Mediterranean Conference on Pattern Recognition and Artificial Intelligence, pp. 88\u201394 (2018)","DOI":"10.1145\/3177148.3180092"},{"key":"3202_CR28","doi-asserted-by":"crossref","unstructured":"Slimani, K., Lekdioui, K., Messoussi, R., Touahni, R.: Compound facial expression recognition based on highway cnn. In: Proceedings of the new challenges in data sciences: acts of the second conference of the Moroccan Classification Society, pp. 1\u20137 (2019)","DOI":"10.1145\/3314074.3314075"},{"key":"3202_CR29","doi-asserted-by":"crossref","unstructured":"Zook, A.: \u201cGame agi beyond characters,\u201d Artificial Intelligence: Concepts, Methodologies, Tools, and Applications, pp. 463\u2013484, (2017)","DOI":"10.4018\/978-1-5225-1759-7.ch020"},{"key":"3202_CR30","doi-asserted-by":"crossref","unstructured":"Everitt, T., Lea, G., Hutter, M.: Agi safety literature review. arXiv preprint arXiv:1805.01109, (2018)","DOI":"10.24963\/ijcai.2018\/768"},{"key":"3202_CR31","doi-asserted-by":"crossref","unstructured":"Hartholt, A., Traum, D., Marsella, S. C., Shapiro, A., Stratou, G., Leuski, A., Morency, L.-P., Gratch, J.: All together now: Introducing the virtual human toolkit. In: Intelligent Virtual Agents: 13th International Conference, IVA: Edinburgh, UK, August 29\u201331, 2013. Proceedings 13. Springer 2013, 368\u2013381 (2013)","DOI":"10.1007\/978-3-642-40415-3_33"},{"key":"3202_CR32","unstructured":"Zhou, W., Peng, X., Riedl, M.: Dialogue shaping: Empowering agents through npc interaction. arXiv preprint arXiv:2307.15833, (2023)"},{"key":"3202_CR33","doi-asserted-by":"crossref","unstructured":"Gong, R., Huang, Q., Ma, X., Vo, H., Durante, Z., Noda, Y., Zheng, Z., Zhu, S.-C., Terzopoulos, D., Fei-Fei, L. et\u00a0al.: Mindagent: Emergent gaming interaction. arXiv preprint arXiv:2309.09971, (2023)","DOI":"10.18653\/v1\/2024.findings-naacl.200"},{"key":"3202_CR34","unstructured":"Csepregi, L.\u00a0M.: \u201cThe effect of context-aware llm-based npc conversations on player engagement in role-playing video games,\u201d AALBORG UNIVERSITY,Student assignment : Master\u2019s thesis and HD graduation project, (2023)"},{"key":"3202_CR35","doi-asserted-by":"crossref","unstructured":"Lee, J., Marsella, S.: Nonverbal behavior generator for embodied conversational agents. In: International Workshop on Intelligent Virtual Agents. Springer, pp. 243\u2013255 (2006)","DOI":"10.1007\/11821830_20"},{"key":"3202_CR36","doi-asserted-by":"crossref","unstructured":"Vilhj\u00e1lmsson, H., Cantelmo, N., Cassell, J., Chafai, N.\u00a0E., Kipp, M., Kopp, S., Mancini, M., Marsella, S., Marshall, A.\u00a0N., Pelachaud, C. et\u00a0al.: The behavior markup language: Recent developments and challenges. In: Intelligent Virtual Agents: 7th International Conference, IVA 2007 Paris, France, September 17-19, 2007 Proceedings 7. Springer, pp. 99\u2013111 (2007)","DOI":"10.1007\/978-3-540-74997-4_10"},{"key":"3202_CR37","doi-asserted-by":"crossref","unstructured":"Hartholt, A., Fast, E., Li, Z., Kim, K., Leeds, A., Mozgai, S.:\u201cRe-architecting the virtual human toolkit: towards an interoperable platform for embodied conversational agent research and development,\u201d In: Proceedings of the 22nd ACM International Conference on Intelligent Virtual Agents, pp. 1\u20138 (2022)","DOI":"10.1145\/3514197.3549671"},{"issue":"3","key":"3202_CR38","doi-asserted-by":"publisher","first-page":"382","DOI":"10.1109\/TAFFC.2016.2545650","volume":"8","author":"A Cerekovic","year":"2016","unstructured":"Cerekovic, A., Aran, O., Gatica-Perez, D.: Rapport with virtual agents: What do human social cues and personality explain? IEEE Trans. Affect. Comput. 8(3), 382\u2013395 (2016)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"6","key":"3202_CR39","doi-asserted-by":"publisher","first-page":"1293","DOI":"10.1007\/s12369-020-00680-7","volume":"12","author":"K Loveys","year":"2020","unstructured":"Loveys, K., Sebaratnam, G., Sagar, M., Broadbent, E.: The effect of design features on relationship quality with embodied conversational agents: a systematic review. Int. J. Soc. Robot. 12(6), 1293\u20131312 (2020)","journal-title":"Int. J. Soc. Robot."},{"key":"3202_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2020.102409","volume":"138","author":"S Ter Stal","year":"2020","unstructured":"Ter Stal, S., Kramer, L.L., Tabak, M., Akker, H.. op den, Hermens, H..: Design features of embodied conversational agents in ehealth: a literature review. Int. J. Hum Comput Stud. 138, 102409 (2020)","journal-title":"Int. J. Hum Comput Stud."},{"key":"3202_CR41","doi-asserted-by":"publisher","first-page":"317","DOI":"10.1016\/j.patrec.2019.11.035","volume":"129","author":"D Tellols","year":"2020","unstructured":"Tellols, D., Lopez-Sanchez, M., Rodr\u00edguez, I., Almajano, P., Puig, A.: Enhancing sentient embodied conversational agents with machine learning. Pattern Recogn. Lett. 129, 317\u2013323 (2020)","journal-title":"Pattern Recogn. Lett."},{"key":"3202_CR42","unstructured":"Liu, C., Li, X.: Superimposition-guided facial reconstruction from skull. arXiv preprint arXiv:1810.00107, (2018)"},{"key":"3202_CR43","doi-asserted-by":"crossref","unstructured":"Lattas, A., Moschoglou, S., Gecer, B., Ploumpis, S., Triantafyllou, V., Ghosh, A., Zafeiriou, S.: Avatarme: Realistically renderable 3d facial reconstruction\" in-the-wild. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 760\u2013769 (2020)","DOI":"10.1109\/CVPR42600.2020.00084"},{"key":"3202_CR44","doi-asserted-by":"crossref","unstructured":"Taherkhani, F., Rai, A., Gao, Q., Srivastava, S., Chen, X., de\u00a0la Torre, F., Song, S., Prakash, A., Kim, D.: Controllable 3d generative adversarial face model via disentangling shape and appearance. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 826\u2013836 (2023)","DOI":"10.1109\/WACV56688.2023.00089"},{"key":"3202_CR45","doi-asserted-by":"crossref","unstructured":"Otberdout, N., Ferrari, C., Daoudi, M., Berretti, S., Del\u00a0Bimbo, A.: Sparse to dense dynamic 3d facial expression generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20\u00a0385\u201320\u00a0394 (2022)","DOI":"10.1109\/CVPR52688.2022.01974"},{"key":"3202_CR46","unstructured":"Tikka, P., Bastamow, T., Gerolin, I., Gerry, L.\u00a0J., Pardinho, V.\u00a0P., Kaipainen, M., Kosunen, I.: The state of darkness: non-human narratives embedded in the encounters with artificial agents. In: Non-Human Narratives. CARNEGIE MELLON UNIVERSITY, p.\u00a09 (2020)"},{"key":"3202_CR47","unstructured":"Lugaresi, C., Tang, J., Nash, H., McClanahan, C., Uboweja, E., Hays, M., Zhang, F., Chang, C.-L., Yong, M.\u00a0G., Lee, J. et\u00a0al.: Mediapipe: A framework for building perception pipelines. arXiv preprint arXiv:1906.08172, (2019)"},{"key":"3202_CR48","first-page":"1755","volume":"10","author":"DE King","year":"2009","unstructured":"King, D.E.: Dlib-ml: a machine learning toolkit. J. Mach. Learn. Res. 10, 1755\u20131758 (2009)","journal-title":"J. Mach. Learn. Res."},{"key":"3202_CR49","doi-asserted-by":"publisher","unstructured":"Lang, S., Murrow, G.: The Distance Formula. New York, NY: Springer New York, pp. 110\u2013122. [Online]. Available: https:\/\/doi.org\/10.1007\/978-1-4757-2022-8_4(1988)","DOI":"10.1007\/978-1-4757-2022-8_4"},{"issue":"4","key":"3202_CR50","doi-asserted-by":"publisher","first-page":"193","DOI":"10.2307\/3029196","volume":"31","author":"PC Hammer","year":"1958","unstructured":"Hammer, P.C.: The midpoint method of numerical integration. Math. Mag. 31(4), 193\u2013195 (1958)","journal-title":"Math. Mag."},{"key":"3202_CR51","unstructured":"Mar 1964. [Online]. Available: https:\/\/gis.stackexchange.com\/questions\/256738\/bearing-of-two-close-lines-in-python"},{"key":"3202_CR52","unstructured":"OpenCV, Perspective-n-point (pnp) pose computation. [Online]. Available: https:\/\/docs.opencv.org\/4.x\/d5\/d1f\/calib3d_solvePnP.html"},{"key":"3202_CR53","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen, W.\u00a0V.: Facial action coding system. Environmental Psychology & Nonverbal Behavior, (1978)","DOI":"10.1037\/t27734-000"},{"key":"3202_CR54","doi-asserted-by":"crossref","unstructured":"Baltru\u0161aitis, T., Robinson, P., Morency, L.-P.: Openface: an open source facial behavior analysis toolkit. In: IEEE winter conference on applications of computer vision (WACV). IEEE 2016, 1\u201310 (2016)","DOI":"10.1109\/WACV.2016.7477553"},{"issue":"1","key":"3202_CR55","doi-asserted-by":"publisher","first-page":"458","DOI":"10.3390\/s23010458","volume":"23","author":"AH Sham","year":"2023","unstructured":"Sham, A.H., Khan, A., Lamas, D., Tikka, P., Anbarjafari, G.: Towards context-aware facial emotion reaction database for dyadic interaction settings. Sensors 23(1), 458 (2023)","journal-title":"Sensors"},{"issue":"2","key":"3202_CR56","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1007\/s11760-022-02257-5","volume":"17","author":"AH Sham","year":"2023","unstructured":"Sham, A.H., Tikka, P., Lamas, D., Anbarjafari, G.: Automatic reaction emotion estimation in a human-human dyadic setting using deep neural networks. SIViP 17(2), 527\u2013534 (2023)","journal-title":"SIViP"},{"key":"3202_CR57","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.\u00a0N., Kaiser, L.\u00a0u., Polosukhin, I.: \u201cAttention is all you need,\u201d In: Advances in Neural Information Processing Systems, I.\u00a0Guyon, U.\u00a0V. Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett, Eds., vol.\u00a030. Curran Associates, Inc., 2017. [Online]. Available: https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2017\/file\/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf"},{"key":"3202_CR58","unstructured":"Team, K.: Keras documentation: Timedistributed layer. [Online]. Available: https:\/\/keras.io\/api\/layers\/recurrent_layers\/time_distributed\/"},{"key":"3202_CR59","first-page":"59","volume":"1","author":"JF Kenney","year":"1962","unstructured":"Kenney, J.F., Keeping, E.: Root mean square. Math. Stat. 1, 59\u201360 (1962)","journal-title":"Math. Stat."},{"key":"3202_CR60","unstructured":"Reaction facial expression synthesis on an avatar from one person\u2019s facial expressions. Jul 2023. [Online]. Available: https:\/\/youtu.be\/bVLZlRqlY7k"},{"key":"3202_CR61","unstructured":"Facial mimicry on an avatar from one person\u2019s facial expressions using openface api. Jul 2023. [Online]. Available: https:\/\/youtu.be\/H6h_JpLNtyA"},{"key":"3202_CR62","unstructured":"Facial mimicry using google\u2019s mediapipe. Jul 2023. [Online]. Available: https:\/\/youtu.be\/9i1Ntrg5E28"},{"issue":"6","key":"3202_CR63","doi-asserted-by":"publisher","first-page":"194","DOI":"10.1145\/3130800.3130813","volume":"36","author":"T Li","year":"2017","unstructured":"Li, T., Bolkart, T., Black, M.J., Li, H., Romero, J.: Learning a model of facial shape and expression from 4d scans. ACM Trans. Graph. 36(6), 194\u20131 (2017)","journal-title":"ACM Trans. Graph."}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03202-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03202-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03202-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,18]],"date-time":"2024-11-18T19:19:12Z","timestamp":1731957552000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03202-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,11]]},"references-count":63,"journal-issue":{"issue":"S1","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["3202"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03202-4","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2024,5,11]]},"assertion":[{"value":"18 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 March 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 April 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}