{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,2]],"date-time":"2025-10-02T00:18:14Z","timestamp":1759364294824,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032027245","type":"print"},{"value":"9783032027252","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:00:00Z","timestamp":1759276800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:00:00Z","timestamp":1759276800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-02725-2_36","type":"book-chapter","created":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:47:57Z","timestamp":1759279677000},"page":"459-470","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Expression Recognition in\u00a0Faces Partially Occluded by\u00a0Head-Mounted Displays"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3153-2088","authenticated-orcid":false,"given":"Jos\u00e9 L.","family":"G\u00f3mez-Sirvent","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1407-9886","authenticated-orcid":false,"given":"Francisco L\u00f3pez de la","family":"Rosa","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4455-370X","authenticated-orcid":false,"given":"Roberto","family":"S\u00e1nchez-Reolid","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8211-0398","authenticated-orcid":false,"given":"Antonio","family":"Fern\u00e1ndez-Caballero","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,1]]},"reference":[{"key":"36_CR1","doi-asserted-by":"publisher","first-page":"388","DOI":"10.1016\/j.procs.2019.08.230","volume":"157","author":"MT Akbar","year":"2019","unstructured":"Akbar, M.T., Ilmi, M.N., Rumayar, I.V., Moniaga, J., Chen, T.K., Chowanda, A.: Enhancing game experience with facial expression recognition as dynamic balancing. Procedia Comput. Sci. 157, 388\u2013395 (2019). https:\/\/doi.org\/10.1016\/j.procs.2019.08.230","journal-title":"Procedia Comput. Sci."},{"key":"36_CR2","doi-asserted-by":"publisher","first-page":"44613","DOI":"10.1109\/access.2024.3380439","volume":"12","author":"A Casas-Ortiz","year":"2024","unstructured":"Casas-Ortiz, A., Echeverria, J., Jimenez-Tellez, N., Santos, O.C.: Exploring the impact of partial occlusion on emotion classification from facial expressions: a comparative study of XR headsets and face masks. IEEE Access 12, 44613\u201344627 (2024). https:\/\/doi.org\/10.1109\/access.2024.3380439","journal-title":"IEEE Access"},{"issue":"3","key":"36_CR3","doi-asserted-by":"publisher","first-page":"1717","DOI":"10.1007\/s10055-022-00720-9","volume":"27","author":"X Chen","year":"2022","unstructured":"Chen, X., Chen, H.: Emotion recognition using facial expressions in an immersive virtual reality application. Virtual Reality 27(3), 1717\u20131732 (2022). https:\/\/doi.org\/10.1007\/s10055-022-00720-9","journal-title":"Virtual Reality"},{"key":"36_CR4","doi-asserted-by":"publisher","unstructured":"ELsayed, Y., ELSayed, A., Abdou, M.A.: An automatic improved facial expression recognition for masked faces. Neural Comput. Appl. 35(20), 14963\u201314972 (2023). https:\/\/doi.org\/10.1007\/s00521-023-08498-w","DOI":"10.1007\/s00521-023-08498-w"},{"key":"36_CR5","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: Slowfast networks for video recognition (2018)","DOI":"10.1109\/ICCV.2019.00630"},{"key":"36_CR6","doi-asserted-by":"publisher","unstructured":"G\u00f3mez-Sirvent, J.L., L\u00f3pez de\u00a0la Rosa, F., L\u00f3pez, M.T., Fern\u00e1ndez-Caballero, A.: Facial expression recognition in the wild for low-resolution images using voting residual network. Electronics 12(18), 3837 (2023). https:\/\/doi.org\/10.3390\/electronics12183837","DOI":"10.3390\/electronics12183837"},{"key":"36_CR7","doi-asserted-by":"publisher","unstructured":"Hasani, B., Mahoor, M.H.: Facial expression recognition using enhanced deep 3D convolutional neural networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 2278\u20132288. IEEE (2017). https:\/\/doi.org\/10.1109\/cvprw.2017.282","DOI":"10.1109\/cvprw.2017.282"},{"key":"36_CR8","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners (2021)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"36_CR9","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"16","key":"36_CR10","doi-asserted-by":"publisher","first-page":"2181","DOI":"10.1016\/j.patrec.2012.07.015","volume":"33","author":"X Huang","year":"2012","unstructured":"Huang, X., Zhao, G., Zheng, W., Pietik\u00e4inen, M.: Towards a dynamic expression recognition system under facial occlusion. Pattern Recogn. Lett. 33(16), 2181\u20132191 (2012). https:\/\/doi.org\/10.1016\/j.patrec.2012.07.015","journal-title":"Pattern Recogn. Lett."},{"key":"36_CR11","doi-asserted-by":"publisher","unstructured":"Huang, Y., Peng, J., Cai, Z., Guo, J., Chen, G., Tan, S.: Facial expression recognition with age-group expression feature learning. In: 2024 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20138. IEEE (2024). https:\/\/doi.org\/10.1109\/ijcnn60899.2024.10649944","DOI":"10.1109\/ijcnn60899.2024.10649944"},{"key":"36_CR12","doi-asserted-by":"publisher","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The Extended Cohn-Kanade Dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops. IEEE (2010). https:\/\/doi.org\/10.1109\/cvprw.2010.5543262","DOI":"10.1109\/cvprw.2010.5543262"},{"key":"36_CR13","unstructured":"Lugaresi, C., et al.: MediaPipe: a framework for building perception pipelines (2019)"},{"key":"36_CR14","doi-asserted-by":"publisher","unstructured":"Melaugh, R., Siddique, N., Coleman, S., Yogarajah, P.: Facial expression recognition on partial facial sections. In: 2019 11th International Symposium on Image and Signal Processing and Analysis (ISPA), pp. 193\u2013197. IEEE (2019). https:\/\/doi.org\/10.1109\/ispa.2019.8868630","DOI":"10.1109\/ispa.2019.8868630"},{"key":"36_CR15","doi-asserted-by":"crossref","unstructured":"Ortmann, T., Wang, Q., Putzar, L.: EmojiHeroVR: a study on facial expression recognition under partial occlusion from head-mounted displays (2024)","DOI":"10.1109\/ACII63134.2024.00014"},{"key":"36_CR16","doi-asserted-by":"publisher","first-page":"446","DOI":"10.1109\/tip.2021.3129120","volume":"31","author":"D Poux","year":"2022","unstructured":"Poux, D., Allaert, B., Ihaddadene, N., Bilasco, I.M., Djeraba, C., Bennamoun, M.: Dynamic facial expression recognition under partial occlusion with optical flow reconstruction. IEEE Trans. Image Process. 31, 446\u2013457 (2022). https:\/\/doi.org\/10.1109\/tip.2021.3129120","journal-title":"IEEE Trans. Image Process."},{"issue":"15","key":"36_CR17","doi-asserted-by":"publisher","first-page":"22405","DOI":"10.1007\/s11042-020-08993-5","volume":"80","author":"D Poux","year":"2020","unstructured":"Poux, D., Allaert, B., Mennesson, J., Ihaddadene, N., Bilasco, I.M., Djeraba, C.: Facial expressions analysis under occlusions based on specificities of facial motion propagation. Multimedia Tools Appl. 80(15), 22405\u201322427 (2020). https:\/\/doi.org\/10.1007\/s11042-020-08993-5","journal-title":"Multimedia Tools Appl."},{"issue":"11","key":"36_CR18","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0188062","volume":"12","author":"T Pozzoli","year":"2017","unstructured":"Pozzoli, T., Gini, G., Alto\u00e8, G.: Associations between facial emotion recognition and young adolescents\u2019 behaviors in bullying. PLoS ONE 12(11), e0188062 (2017). https:\/\/doi.org\/10.1371\/journal.pone.0188062","journal-title":"PLoS ONE"},{"key":"36_CR19","doi-asserted-by":"publisher","unstructured":"Rodrigues, A.S.F., Lopes, J.C., Lopes, R.P., Teixeira, L.F.: Classification of facial expressions under partial occlusion for VR games. In: Optimization, Learning Algorithms and Applications, pp. 804\u2013819. Springer International Publishing (2022). https:\/\/doi.org\/10.1007\/978-3-031-23236-7_55","DOI":"10.1007\/978-3-031-23236-7_55"},{"issue":"1","key":"36_CR20","doi-asserted-by":"publisher","first-page":"1057","DOI":"10.1109\/tetci.2023.3332891","volume":"8","author":"HA Shehu","year":"2024","unstructured":"Shehu, H.A., Browne, W.N., Eisenbarth, H.: Attention-based methods for emotion categorization from partially covered faces. IEEE Trans. Emerging Top. Comput. Intell. 8(1), 1057\u20131070 (2024). https:\/\/doi.org\/10.1109\/tetci.2023.3332891","journal-title":"IEEE Trans. Emerging Top. Comput. Intell."},{"key":"36_CR21","doi-asserted-by":"publisher","unstructured":"Siddiqi, M.H., Ahmad, I., Alhwaiti, Y., Khan, F.: Facial expression recognition for healthcare monitoring systems using neural random forest. IEEE J. Biomed. Health Inf., 1\u201313 (2024). https:\/\/doi.org\/10.1109\/jbhi.2024.3482450","DOI":"10.1109\/jbhi.2024.3482450"},{"key":"36_CR22","doi-asserted-by":"publisher","unstructured":"Sudha, S., Suganya, S.: On-road driver facial expression emotion recognition with parallel multi-verse optimizer (PMVO) and optical flow reconstruction for partial occlusion in internet of things (IoT). Measurement: Sensors 26, 100711 (2023). https:\/\/doi.org\/10.1016\/j.measen.2023.100711","DOI":"10.1016\/j.measen.2023.100711"},{"key":"36_CR23","doi-asserted-by":"publisher","unstructured":"Tang, X., Gong, Y., Xiao, Y., Xiong, J., Bao, L.: Facial expression recognition for probing students\u2019 emotional engagement in science learning. J. Sci. Educ. Technol. (2024). https:\/\/doi.org\/10.1007\/s10956-024-10143-7","DOI":"10.1007\/s10956-024-10143-7"},{"key":"36_CR24","doi-asserted-by":"publisher","unstructured":"Tang, Z., Zhao, Y., Wen, Y., Liu, M.: A survey on backbones for deep video action recognition. In: 2024 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), pp. 1\u20136. IEEE (2024). https:\/\/doi.org\/10.1109\/icmew63481.2024.10645423","DOI":"10.1109\/icmew63481.2024.10645423"},{"key":"36_CR25","unstructured":"Tong, Z., Song, Y., Wang, J., Wang, L.: Videomae: masked autoencoders are data-efficient learners for self-supervised video pre-training (2022)"},{"key":"36_CR26","doi-asserted-by":"publisher","unstructured":"Wang, D., Gu, Y., Luo, L., Ren, F.: Occlusion-aware visual-language model for occluded facial expression recognition. In: 2024 International Joint Conference on Neural Networks (IJCNN). pp. 1\u20138. IEEE (2024). https:\/\/doi.org\/10.1109\/ijcnn60899.2024.10651502","DOI":"10.1109\/ijcnn60899.2024.10651502"},{"key":"36_CR27","doi-asserted-by":"publisher","unstructured":"Yang, B., Jianming, W., Hattori, G.: Face mask aware robust facial expression recognition during the Covid-19 pandemic. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 240\u2013244. IEEE (2021). https:\/\/doi.org\/10.1109\/icip42928.2021.9506047","DOI":"10.1109\/icip42928.2021.9506047"},{"key":"36_CR28","doi-asserted-by":"publisher","first-page":"173","DOI":"10.1016\/j.patrec.2022.11.004","volume":"164","author":"B Yang","year":"2022","unstructured":"Yang, B., et al.: Face-mask-aware facial expression recognition based on face parsing and vision transformer. Pattern Recogn. Lett. 164, 173\u2013182 (2022). https:\/\/doi.org\/10.1016\/j.patrec.2022.11.004","journal-title":"Pattern Recogn. Lett."},{"key":"36_CR29","doi-asserted-by":"publisher","unstructured":"Yang, X., Liu, Y., Liu, J., Sun, Y.: Hzs-nas: Neural architecture search with hybrid zero-shot proxy for facial expression recognition. In: 2024 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20138. IEEE (2024). https:\/\/doi.org\/10.1109\/ijcnn60899.2024.10650259","DOI":"10.1109\/ijcnn60899.2024.10650259"},{"issue":"9","key":"36_CR30","doi-asserted-by":"publisher","first-page":"607","DOI":"10.1016\/j.imavis.2011.07.002","volume":"29","author":"G Zhao","year":"2011","unstructured":"Zhao, G., Huang, X., Taini, M., Li, S.Z., Pietik\u00e4inen, M.: Facial expression recognition from near-infrared videos. Image Vis. Comput. 29(9), 607\u2013619 (2011). https:\/\/doi.org\/10.1016\/j.imavis.2011.07.002","journal-title":"Image Vis. Comput."},{"key":"36_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102261","volume":"106","author":"X Zhou","year":"2024","unstructured":"Zhou, X., Li, J., Lan, G., Ni, R., Cangelosi, A., Wang, J., Liu, X.: Efficient lower layers parameter decoupling personalized federated learning method of facial expression recognition for home care robots. Inf. Fusion 106, 102261 (2024). https:\/\/doi.org\/10.1016\/j.inffus.2024.102261","journal-title":"Inf. Fusion"}],"container-title":["Lecture Notes in Computer Science","Advances in Computational Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-02725-2_36","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T00:48:03Z","timestamp":1759279683000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-02725-2_36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,1]]},"ISBN":["9783032027245","9783032027252"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-02725-2_36","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,1]]},"assertion":[{"value":"1 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IWANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Work-Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"A Coru\u00f1a","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Spain","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 June 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 June 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iwann2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iwann.uma.es\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}