{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:15:42Z","timestamp":1757618142959,"version":"3.44.0"},"publisher-location":"Cham","reference-count":71,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031915802"},{"type":"electronic","value":"9783031915819"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-91581-9_29","type":"book-chapter","created":{"date-parts":[[2025,5,27]],"date-time":"2025-05-27T11:22:25Z","timestamp":1748344945000},"page":"404-421","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Ig3D: Integrating 3D Face Representations in\u00a0Facial Expression Inference"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-4036-7690","authenticated-orcid":false,"given":"Lu","family":"Dong","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-0511-8495","authenticated-orcid":false,"given":"Xiao","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7118-9280","authenticated-orcid":false,"given":"Srirangaraj","family":"Setlur","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5318-7409","authenticated-orcid":false,"given":"Venu","family":"Govindaraju","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1414-6433","authenticated-orcid":false,"given":"Ifeoma","family":"Nwogu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"key":"29_CR1","doi-asserted-by":"crossref","unstructured":"Antoniadis, P., Pikoulis, I., Filntisis, P.P., Maragos, P.: An audiovisual and contextual approach for categorical and continuous emotion recognition in-the-wild. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3645\u20133651 (2021)","DOI":"10.1109\/ICCVW54120.2021.00407"},{"issue":"1","key":"29_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1177\/1529100619832930","volume":"20","author":"LF Barrett","year":"2019","unstructured":"Barrett, L.F., Adolphs, R., Marsella, S., Martinez, A.M., Pollak, S.D.: Emotional expressions reconsidered: challenges to inferring emotion from human facial movements. Psychological Sci. Public Interest 20(1), 1\u201368 (2019)","journal-title":"Psychological Sci. Public Interest"},{"key":"29_CR3","doi-asserted-by":"publisher","unstructured":"Benitez-Quiroz, C.F., Srinivasan, R., Martinez, A.M.: Emotionet: a large-scale, fine-grained dataset for automatic facial expression analysis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5562\u20135570 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00567, https:\/\/ieeexplore.ieee.org\/document\/9156975","DOI":"10.1109\/CVPR42600.2020.00567"},{"issue":"3","key":"29_CR4","doi-asserted-by":"publisher","first-page":"329","DOI":"10.1049\/iet-cvi.2018.5281","volume":"13","author":"C Bian","year":"2019","unstructured":"Bian, C., Zhang, Y., Yang, F., Bi, W., Lu, W.: Spontaneous facial expression database for academic emotion inference in online learning. IET Comput. Vision 13(3), 329\u2013337 (2019)","journal-title":"IET Comput. Vision"},{"key":"29_CR5","doi-asserted-by":"crossref","unstructured":"Bulat, A., Cheng, S., Yang, J., Garbett, A., Sanchez, E., Tzimiropoulos, G.: Pre-training strategies and datasets for facial representation learning. In: European Conference on Computer Vision, pp. 107\u2013125. Springer (2022)","DOI":"10.1007\/978-3-031-19778-9_7"},{"issue":"1","key":"29_CR6","doi-asserted-by":"publisher","first-page":"109","DOI":"10.3758\/BRM.40.1.109","volume":"40","author":"MG Calvo","year":"2008","unstructured":"Calvo, M.G., Lundqvist, D.: Facial expressions of emotion (kdef): identification under different display-duration conditions. Behav. Res. Methods 40(1), 109\u2013115 (2008)","journal-title":"Behav. Res. Methods"},{"issue":"4","key":"29_CR7","doi-asserted-by":"publisher","first-page":"1906","DOI":"10.1109\/TAFFC.2022.3201290","volume":"13","author":"K Chen","year":"2022","unstructured":"Chen, K., Yang, X., Fan, C., Zhang, W., Ding, Y.: Semantic-rich facial emotional expression recognition. IEEE Trans. Affect. Comput. 13(4), 1906\u20131916 (2022)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"29_CR8","unstructured":"Chen, Y., Li, J., Shan, S., Wang, M., Hong, R.: From static to dynamic: adapting landmark-aware image models for facial expression recognition in videos. arXiv preprint arXiv:2312.05447 (2023)"},{"key":"29_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2022.104569","volume":"128","author":"Z Chen","year":"2022","unstructured":"Chen, Z., Feng, X., Zhang, S.: Emotion detection and face recognition of drivers in autonomous vehicles in iot platform. Image Vis. Comput. 128, 104569 (2022)","journal-title":"Image Vis. Comput."},{"key":"29_CR10","doi-asserted-by":"crossref","unstructured":"Dan\u011b\u010dek, R., Black, M.J., Bolkart, T.: Emoca: emotion driven monocular face capture and animation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20311\u201320322 (2022)","DOI":"10.1109\/CVPR52688.2022.01967"},{"issue":"11","key":"29_CR11","doi-asserted-by":"publisher","first-page":"2049","DOI":"10.1109\/TMM.2015.2477042","volume":"17","author":"C Ding","year":"2015","unstructured":"Ding, C., Tao, D.: Robust face recognition via multimodal deep face representation. IEEE Trans. Multimedia 17(11), 2049\u20132058 (2015)","journal-title":"IEEE Trans. Multimedia"},{"key":"29_CR12","doi-asserted-by":"crossref","unstructured":"Dong, L., Chaudhary, L., Xu, F., Wang, X., Lary, M., Nwogu, I.: Signavatar: Sign language 3d motion reconstruction and generation. arXiv preprint arXiv:2405.07974 (2024)","DOI":"10.1109\/FG59268.2024.10581934"},{"issue":"3\u20134","key":"29_CR13","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman, P.: An argument for basic emotions. Cognition Emotion 6(3\u20134), 169\u2013200 (1992)","journal-title":"Cognition Emotion"},{"key":"29_CR14","unstructured":"Epstein, S.: Controversial issues in emotion theory. Review of Personality & Social Psychology (1984)"},{"issue":"4","key":"29_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459936","volume":"40","author":"Y Feng","year":"2021","unstructured":"Feng, Y., Feng, H., Black, M.J., Bolkart, T.: Learning an animatable detailed 3d face model from in-the-wild images. ACM Trans. Graph. (ToG) 40(4), 1\u201313 (2021)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"29_CR16","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/j.neunet.2014.09.005","volume":"64","author":"IJ Goodfellow","year":"2015","unstructured":"Goodfellow, I.J.: Challenges in representation learning: a report on three machine learning contests. Neural Netw. 64, 59\u201363 (2015)","journal-title":"Neural Netw."},{"issue":"1","key":"29_CR17","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1109\/5.554205","volume":"85","author":"DL Hall","year":"1997","unstructured":"Hall, D.L., Llinas, J.: An introduction to multisensor data fusion. Proc. IEEE 85(1), 6\u201323 (1997)","journal-title":"Proc. IEEE"},{"key":"29_CR18","unstructured":"Her, M.B., Jeong, J., Song, H., Han, J.H.: Batch transformer: look for attention in batch. arXiv preprint arXiv:2407.04218 (2024)"},{"key":"29_CR19","doi-asserted-by":"crossref","unstructured":"Izard, C.E.: Basic emotions, relations among emotions, and emotion-cognition relations (1992)","DOI":"10.1007\/978-1-4899-0615-1_3"},{"key":"29_CR20","doi-asserted-by":"crossref","unstructured":"Kahou, S.E., et\u00a0al.: Combining modality specific deep neural networks for emotion recognition in video. In: Proceedings of the 15th ACM on International Conference on Multimodal Interaction, pp. 543\u2013550 (2013)","DOI":"10.1145\/2522848.2531745"},{"key":"29_CR21","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1016\/j.imavis.2019.02.004","volume":"83","author":"RA Khan","year":"2019","unstructured":"Khan, R.A., Crenn, A., Meyer, A., Bouakaz, S.: A novel database of children\u2019s spontaneous facial expressions (liris-cse). Image Vis. Comput. 83, 61\u201369 (2019)","journal-title":"Image Vis. Comput."},{"key":"29_CR22","doi-asserted-by":"crossref","unstructured":"Kollias, D., Schulc, A., Hajiyev, E., Zafeiriou, S.: Analysing affective behavior in the first abaw 2020 competition. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020) (FG), pp. 794\u2013800 (2020)","DOI":"10.1109\/FG47880.2020.00126"},{"key":"29_CR23","doi-asserted-by":"crossref","unstructured":"Kollias, D.: Abaw: Valence-arousal estimation, expression recognition, action unit detection & multi-task learning challenges. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2328\u20132336 (2022)","DOI":"10.1109\/CVPRW56347.2022.00259"},{"key":"29_CR24","doi-asserted-by":"crossref","unstructured":"Kollias, D.: Abaw: learning from synthetic data & multi-task learning challenges. In: European Conference on Computer Vision, pp. 157\u2013172. Springer (2023)","DOI":"10.1007\/978-3-031-25075-0_12"},{"key":"29_CR25","doi-asserted-by":"crossref","unstructured":"Kollias, D.: Multi-label compound expression recognition: C-expr database & network. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5589\u20135598 (2023)","DOI":"10.1109\/CVPR52729.2023.00541"},{"key":"29_CR26","unstructured":"Kollias, D., Sharmanska, V., Zafeiriou, S.: Face behavior a la carte: Expressions, affect and action units in a single network. arXiv preprint arXiv:1910.11111 (2019)"},{"key":"29_CR27","unstructured":"Kollias, D., Sharmanska, V., Zafeiriou, S.: Distribution matching for heterogeneous multi-task learning: a large-scale face study. arXiv preprint arXiv:2105.03790 (2021)"},{"key":"29_CR28","doi-asserted-by":"crossref","unstructured":"Kollias, D., Tzirakis, P., Baird, A., Cowen, A., Zafeiriou, S.: Abaw: valence-arousal estimation, expression recognition, action unit detection & emotional reaction intensity estimation challenges. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5888\u20135897 (2023)","DOI":"10.1109\/CVPRW59228.2023.00626"},{"key":"29_CR29","doi-asserted-by":"crossref","unstructured":"Kollias, D., et al.: The 6th affective behavior analysis in-the-wild (abaw) competition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4587\u20134598 (2024)","DOI":"10.1109\/CVPRW63382.2024.00461"},{"key":"29_CR30","unstructured":"Kollias, D., et al.: Deep affect prediction in-the-wild: Aff-wild database and challenge, deep architectures, and beyond. International Journal of Computer Vision, pp. 1\u201323 (2019)"},{"key":"29_CR31","unstructured":"Kollias, D., Zafeiriou, S.: Expression, affect, action unit recognition: Aff-wild2, multi-task learning and arcface. arXiv preprint arXiv:1910.04855 (2019)"},{"key":"29_CR32","unstructured":"Kollias, D., Zafeiriou, S.: Affect analysis in-the-wild: Valence-arousal, expressions, action units and a unified framework. arXiv preprint arXiv:2103.15792 (2021)"},{"key":"29_CR33","doi-asserted-by":"crossref","unstructured":"Kollias, D., Zafeiriou, S.: Analysing affective behavior in the second abaw2 competition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3652\u20133660 (2021)","DOI":"10.1109\/ICCVW54120.2021.00408"},{"key":"29_CR34","doi-asserted-by":"crossref","unstructured":"Kollias, D., Zafeiriou, S., Kotsia, I., Dhall, A., Ghosh, S., Shao, C., Hu, G.: 7th abaw competition: Multi-task learning and compound expression recognition. arXiv preprint arXiv:2407.03835 (2024)","DOI":"10.1007\/978-3-031-91581-9_3"},{"key":"29_CR35","doi-asserted-by":"publisher","unstructured":"Kosti, R., Alvarez, J.M., Recasens, A., Lapedriza, A.: Emotion recognition in context based on emotic dataset. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 889\u2013895. IEEE (2017). https:\/\/doi.org\/10.1109\/CVPRW.2017.112, https:\/\/arxiv.org\/abs\/2003.13401","DOI":"10.1109\/CVPRW.2017.112"},{"key":"29_CR36","doi-asserted-by":"crossref","unstructured":"Kuncheva, L.I.: Combining pattern classifiers: methods and algorithms. John Wiley & Sons (2014)","DOI":"10.1002\/9781118914564"},{"issue":"9","key":"29_CR37","doi-asserted-by":"publisher","first-page":"1449","DOI":"10.1109\/JPROC.2015.2460697","volume":"103","author":"D Lahat","year":"2015","unstructured":"Lahat, D., Adali, T., Jutten, C.: Multimodal data fusion: an overview of methods, challenges, and prospects. Proc. IEEE 103(9), 1449\u20131477 (2015)","journal-title":"Proc. IEEE"},{"issue":"1","key":"29_CR38","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/TIP.2018.2868382","volume":"28","author":"S Li","year":"2019","unstructured":"Li, S., Deng, W.: Reliable crowdsourcing and deep locality-preserving learning for unconstrained facial expression recognition. IEEE Trans. Image Process. 28(1), 356\u2013370 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"29_CR39","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2584\u20132593. IEEE (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"29_CR40","doi-asserted-by":"crossref","unstructured":"Li, T., Bolkart, T., Black, M.J., Li, H., Romero, J.: Learning a model of facial shape and expression from 4D scans. ACM Trans. Graph. (Proc. SIGGRAPH Asia) 36(6), 194:1\u2013194:17 (2017). https:\/\/doi.org\/10.1145\/3130800.3130813","DOI":"10.1145\/3130800.3130813"},{"issue":"1","key":"29_CR41","doi-asserted-by":"publisher","first-page":"1411","DOI":"10.1016\/j.aej.2020.10.061","volume":"60","author":"Z Li","year":"2021","unstructured":"Li, Z., Zhang, T., Jing, X., Wang, Y.: Facial expression-based analysis on emotion correlations, hotspots, and potential occurrence of urban crimes. Alex. Eng. J. 60(1), 1411\u20131420 (2021)","journal-title":"Alex. Eng. J."},{"key":"29_CR42","doi-asserted-by":"publisher","unstructured":"Liu, Y., et al.: Mafw: a large-scale, multi-modal, compound affective database for dynamic facial expression recognition in the wild. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3688\u20133697 (2022). https:\/\/doi.org\/10.1109\/ICCVW54120.2022.00108, https:\/\/arxiv.org\/abs\/2208.00847","DOI":"10.1109\/ICCVW54120.2022.00108"},{"key":"29_CR43","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended cohn-kanade dataset (ck+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, pp. 94\u2013101. IEEE (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"29_CR44","unstructured":"Lyons, M.J., Kamachi, M., Gyoba, J.: The Japanese female facial expression (jaffe) database. Zenodo (1998)"},{"issue":"1","key":"29_CR45","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1037\/h0024532","volume":"6","author":"A Mehrabian","year":"1967","unstructured":"Mehrabian, A., Wiener, M.: Decoding of inconsistent communications. J. Pers. Soc. Psychol. 6(1), 109 (1967)","journal-title":"J. Pers. Soc. Psychol."},{"key":"29_CR46","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: Affectnet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affective Comput. (2017)"},{"key":"29_CR47","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2024.108822","volume":"179","author":"M Munsif","year":"2024","unstructured":"Munsif, M., Sajjad, M., Ullah, M., Tarekegn, A.N., Cheikh, F.A., Tsakanikas, P., Muhammad, K.: Optimized efficient attention-based network for facial expressions analysis in neurological health care. Comput. Biol. Med. 179, 108822 (2024)","journal-title":"Comput. Biol. Med."},{"issue":"6","key":"29_CR48","doi-asserted-by":"publisher","first-page":"4435","DOI":"10.1016\/j.aej.2021.09.066","volume":"61","author":"Y Nan","year":"2022","unstructured":"Nan, Y., Ju, J., Hua, Q., Zhang, H., Wang, B.: A-mobilenet: an approach of facial expression recognition. Alex. Eng. J. 61(6), 4435\u20134444 (2022)","journal-title":"Alex. Eng. J."},{"key":"29_CR49","unstructured":"Ning, M., Salah, A.A., Ertugrul, I.O.: Representation learning and identity adversarial training for facial behavior understanding. arXiv preprint arXiv:2407.11243 (2024)"},{"key":"29_CR50","unstructured":"Paper with code (2024). https:\/\/paperswithcode.com\/dataset\/affectnet. Accessed 08 July 2024"},{"issue":"4","key":"29_CR51","doi-asserted-by":"publisher","first-page":"377","DOI":"10.1080\/19368623.2019.1647124","volume":"29","author":"C Prentice","year":"2020","unstructured":"Prentice, C., Dominique Lopes, S., Wang, X.: Emotional intelligence or artificial intelligence-an employee perspective. J. Hospitality Marketing Manage. 29(4), 377\u2013403 (2020)","journal-title":"J. Hospitality Marketing Manage."},{"key":"29_CR52","doi-asserted-by":"crossref","unstructured":"Ren, Z., et al.: Veatic: video-based emotion and affect tracking in context dataset. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 4467\u20134477 (2024)","DOI":"10.1109\/WACV57701.2024.00441"},{"key":"29_CR53","doi-asserted-by":"crossref","unstructured":"Retsinas, G., et al.: : 3d facial expressions through analysis-by-neural-synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2490\u20132501 (2024)","DOI":"10.1109\/CVPR52733.2024.00241"},{"key":"29_CR54","unstructured":"Roseman, I.J.: Cognitive determinants of emotion: a structural theory. Review of personality & social psychology (1984)"},{"issue":"6","key":"29_CR55","doi-asserted-by":"publisher","first-page":"1161","DOI":"10.1037\/h0077714","volume":"39","author":"JA Russell","year":"1980","unstructured":"Russell, J.A.: A circumplex model of affect. J. Pers. Soc. Psychol. 39(6), 1161 (1980)","journal-title":"J. Pers. Soc. Psychol."},{"key":"29_CR56","doi-asserted-by":"crossref","unstructured":"Shan, M., Dong, L., Han, Y., Yao, Y., Liu, T., Nwogu, I., Qi, G.J., Hill, M.: Towards open domain text-driven synthesis of multi-person motions. arXiv preprint arXiv:2405.18483 (2024)","DOI":"10.1007\/978-3-031-73650-6_5"},{"key":"29_CR57","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition. In: Proceedings of the Neural Information Processing Systems (NIPS), vol.\u00a011 (2015)"},{"key":"29_CR58","unstructured":"Tan, M., Le, Q.: Efficientnetv2: Smaller models and faster training. In: International Conference on Machine Learning, pp. 10096\u201310106. PMLR (2021)"},{"issue":"1","key":"29_CR59","doi-asserted-by":"publisher","first-page":"42","DOI":"10.1038\/s42256-020-00280-0","volume":"3","author":"A Toisoul","year":"2021","unstructured":"Toisoul, A., Kossaifi, J., Bulat, A., Tzimiropoulos, G., Pantic, M.: Estimation of continuous valence and arousal levels from faces in naturalistic conditions. Nature Mach. Intell. 3(1), 42\u201350 (2021)","journal-title":"Nature Mach. Intell."},{"key":"29_CR60","doi-asserted-by":"crossref","unstructured":"Tu, Z., Talebi, H., Zhang, H., Yang, F., Milanfar, P., Bovik, A., Li, Y.: Maxvit: multi-axis vision transformer. In: European Conference on Computer Vision, pp. 459\u2013479. Springer (2022)","DOI":"10.1007\/978-3-031-20053-3_27"},{"key":"29_CR61","doi-asserted-by":"crossref","unstructured":"Wagner, N., M\u00e4tzler, F., Vossberg, S.R., Schneider, H., Pavlitska, S., Z\u00f6llner, J.M.: Cage: circumplex affect guided expression inference. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4683\u20134692 (2024)","DOI":"10.1109\/CVPRW63382.2024.00471"},{"key":"29_CR62","doi-asserted-by":"publisher","unstructured":"Wang, J., Li, Z., Liu, Y., Fu, Q.: Semantic-rich facial emotional expression recognition with emo135 dataset. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pp. 1150\u20131157 (2021). https:\/\/doi.org\/10.1609\/aaai.v35i2.16214, https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/16214","DOI":"10.1609\/aaai.v35i2.16214"},{"key":"29_CR63","unstructured":"Wasi, A.T., \u0160erbetar, K., Islam, R., Rafi, T.H., Chae, D.K.: Arbex: Attentive feature extraction with reliability balancing for robust facial expression learning. arXiv preprint arXiv:2305.01486 (2023)"},{"issue":"8","key":"29_CR64","doi-asserted-by":"publisher","first-page":"1583","DOI":"10.1109\/TPAMI.2016.2537340","volume":"38","author":"D Wu","year":"2016","unstructured":"Wu, D., Pigou, L., Kindermans, P.J., Le, N., Shao, L., Dambre, J., Odobez, J.M.: Deep dynamic neural networks for multimodal gesture segmentation and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 38(8), 1583\u20131597 (2016)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"29_CR65","doi-asserted-by":"crossref","unstructured":"Xu, F., Chaudhary, L., Dong, L., Setlur, S., Govindaraju, V., Nwogu, I.: A comparative study of video-based human representations for american sign language alphabet generation. In: 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition (FG), pp.\u00a01\u20136. IEEE (2024)","DOI":"10.1109\/FG59268.2024.10582020"},{"key":"29_CR66","doi-asserted-by":"crossref","unstructured":"Yi, D., Lei, Z., Li, S.Z.: Shared representation learning for heterogenous face recognition. In: 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), vol.\u00a01, pp.\u00a01\u20137. IEEE (2015)","DOI":"10.1109\/FG.2015.7163093"},{"issue":"7","key":"29_CR67","doi-asserted-by":"publisher","first-page":"1306","DOI":"10.1002\/aur.1778","volume":"10","author":"SS Yun","year":"2017","unstructured":"Yun, S.S., Choi, J., Park, S.K., Bong, G.Y., Yoo, H.: Social skills training for children with autism spectrum disorder using a robotic behavioral intervention system. Autism Res. 10(7), 1306\u20131323 (2017)","journal-title":"Autism Res."},{"key":"29_CR68","doi-asserted-by":"crossref","unstructured":"Zafeiriou, S., Kollias, D., Nicolaou, M.A., Papaioannou, A., Zhao, G., Kotsia, I.: Aff-wild: Valence and arousal \u2018in-the-wild\u2019 challenge. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1980\u20131987. IEEE (2017)","DOI":"10.1109\/CVPRW.2017.248"},{"key":"29_CR69","doi-asserted-by":"crossref","unstructured":"Zhai, Y., Huang, M., Luan, T., Dong, L., Nwogu, I., Lyu, S., Doermann, D., Yuan, J.: Language-guided human motion synthesis with atomic actions. In: Proceedings of the 31st ACM International Conference on Multimedia, pp. 5262\u20135271 (2023)","DOI":"10.1145\/3581783.3612289"},{"issue":"17","key":"29_CR70","doi-asserted-by":"publisher","first-page":"3595","DOI":"10.3390\/electronics12173595","volume":"12","author":"S Zhang","year":"2023","unstructured":"Zhang, S., Zhang, Y., Zhang, Y., Wang, Y., Song, Z.: A dual-direction attention mixed feature network for facial expression recognition. Electronics 12(17), 3595 (2023)","journal-title":"Electronics"},{"key":"29_CR71","doi-asserted-by":"publisher","unstructured":"Zhang, W., Liu, Y., Chen, J., Qiu, X.: Fine-grained facial expression database (f2ed) for emotion perception and expression recognition. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW), pp. 1720\u20131728 (2020). https:\/\/doi.org\/10.1109\/ICCVW50498.2020.00204, https:\/\/ieeexplore.ieee.org\/document\/9200887","DOI":"10.1109\/ICCVW50498.2020.00204"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-91581-9_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T15:57:19Z","timestamp":1757174239000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-91581-9_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031915802","9783031915819"],"references-count":71,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-91581-9_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}