{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:21:02Z","timestamp":1763644862103,"version":"3.45.0"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T00:00:00Z","timestamp":1758672000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T00:00:00Z","timestamp":1758672000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Natural Science Research Project of the Education Department of Anhui Province of China","award":["2022AH051038","2024AH051116"],"award-info":[{"award-number":["2022AH051038","2024AH051116"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176084"],"award-info":[{"award-number":["62176084"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00371-025-04190-3","type":"journal-article","created":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T15:34:11Z","timestamp":1758728051000},"page":"12865-12888","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DIR-EC: intra-region enhancement and inter-region collaboration network for facial expression recognition"],"prefix":"10.1007","volume":"41","author":[{"given":"Juan","family":"Liu","sequence":"first","affiliation":[]},{"given":"Ying","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zhong","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Min","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Fuji","family":"Ren","sequence":"additional","affiliation":[]},{"given":"Wen Quan","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,24]]},"reference":[{"key":"4190_CR1","doi-asserted-by":"publisher","first-page":"102330","DOI":"10.1016\/j.displa.2022.102330","volume":"75","author":"N Rathour","year":"2022","unstructured":"Rathour, N., Singh, R., Gehlot, A., Thakur, A., Kumar, A.: The decadal perspective of facial emotion processing and recognition: a survey. Displays 75, 102330 (2022). https:\/\/doi.org\/10.1016\/j.displa.2022.102330","journal-title":"Displays"},{"issue":"5","key":"4190_CR2","doi-asserted-by":"publisher","first-page":"122784.1","DOI":"10.1016\/j.eswa.2023.122784","volume":"242","author":"I Saadi","year":"2024","unstructured":"Saadi, I., Cunningham, D.W., Hillali, H.Y.E.: Driver\u2019s facial expression recognition: a comprehensive survey. Expert Syst. Appl. 242(5), 122784.1-122784.34 (2024). https:\/\/doi.org\/10.1016\/j.eswa.2023.122784","journal-title":"Expert Syst. Appl."},{"key":"4190_CR3","doi-asserted-by":"publisher","first-page":"340","DOI":"10.1016\/j.neucom.2020.06.014","volume":"411","author":"J Li","year":"2020","unstructured":"Li, J., Jin, K., Zhou, D., Kubota, N., Ju, Z.: Attention mechanism-based CNN for facial expression recognition. Neurocomputing 411, 340\u2013350 (2020). https:\/\/doi.org\/10.1016\/j.neucom.2020.06.014","journal-title":"Neurocomputing"},{"key":"4190_CR4","doi-asserted-by":"publisher","first-page":"2035","DOI":"10.1007\/s00371-023-02900-3","volume":"40","author":"H Xia","year":"2024","unstructured":"Xia, H., Lu, L., Song, S.: Feature fusion of multi-granularity and multi-scale for facial expression recognition. Vis. Comput. 40, 2035\u20132047 (2024). https:\/\/doi.org\/10.1007\/s00371-023-02900-3","journal-title":"Vis. Comput."},{"key":"4190_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.measen.2023.100711","volume":"26","author":"SS Sudha","year":"2023","unstructured":"Sudha, S.S., Suganya, S.S.: On-road driver facial expression emotion recognition with parallel multi-verse optimizer (PMVO) and optical flow reconstruction for partial occlusion in internet of things (IoT). Meas. Sens. 26, 100711 (2023). https:\/\/doi.org\/10.1016\/j.measen.2023.100711","journal-title":"Meas. Sens."},{"key":"4190_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2022.106621","volume":"215","author":"H Ge","year":"2022","unstructured":"Ge, H., Zhu, Z., Dai, Y., Wang, B., Wu, X.: Facial expression recognition based on deep learning. Comput. Meth. Prog. Bio. 215, 106621 (2022). https:\/\/doi.org\/10.1016\/j.cmpb.2022.106621","journal-title":"Comput. Meth. Prog. Bio."},{"issue":"2","key":"4190_CR7","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1109\/TAI.2022.3172272","volume":"4","author":"R Wadhawan","year":"2023","unstructured":"Wadhawan, R., Gandhi, T.K.: Landmark-aware and part-based ensemble transfer learning network for static facial expression recognition from images. IEEE Trans. Artif. Intell. 4(2), 349\u2013361 (2023). https:\/\/doi.org\/10.1109\/TAI.2022.3172272","journal-title":"IEEE Trans. Artif. Intell."},{"key":"4190_CR8","doi-asserted-by":"publisher","first-page":"370","DOI":"10.1016\/j.ins.2023.02.056","volume":"630","author":"Z Zhang","year":"2023","unstructured":"Zhang, Z., Tian, X., Zhang, Y., Guo, K., Xu, X.: Enhanced discriminative global-local feature learning with priority for facial expression recognition. Inform. Sciences. 630, 370\u2013384 (2023). https:\/\/doi.org\/10.1016\/j.ins.2023.02.056","journal-title":"Inform. Sciences."},{"issue":"9","key":"4190_CR9","doi-asserted-by":"publisher","first-page":"6253","DOI":"10.1109\/TCSVT.2022.3165321","volume":"32","author":"H Liu","year":"2022","unstructured":"Liu, H., Cai, H., Lin, Q., Li, X., Xiao, H.: Adaptive multilayer perceptual attention network for facial expression recognition. IEEE Trans. Circuits Syst. Video Technol. 32(9), 6253\u20136266 (2022). https:\/\/doi.org\/10.1109\/TCSVT.2022.3165321","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"4190_CR10","doi-asserted-by":"publisher","first-page":"58","DOI":"10.1016\/j.patrec.2021.01.029","volume":"145","author":"D Gera","year":"2021","unstructured":"Gera, D., Balasubramanian, S.: Landmark guidance independent spatio-channel attention and complementary context information based facial expression recognition. Pattern Recogn. Lett. 145, 58\u201366 (2021). https:\/\/doi.org\/10.1016\/j.patrec.2021.01.029","journal-title":"Pattern Recogn. Lett."},{"key":"4190_CR11","doi-asserted-by":"publisher","unstructured":"Cai, J., Meng, Z., Khan, A.S., Li, Z., O'Reilly, J., Tong, Y.: Identity-free facial expression recognition using conditional generative adversarial network. In: Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), pp. 19\u201322 (2021). https:\/\/doi.org\/10.1109\/ICIP42928.2021.9506593","DOI":"10.1109\/ICIP42928.2021.9506593"},{"key":"4190_CR12","doi-asserted-by":"publisher","first-page":"4445","DOI":"10.1109\/TIP.2020.2972114","volume":"29","author":"F Zhang","year":"2020","unstructured":"Zhang, F., Zhang, T., Mao, Q., Xu, C.: Geometry guided pose-invariant facial expression recognition. IEEE Trans. Image Process. 29, 4445\u20134460 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2972114","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"4190_CR13","doi-asserted-by":"publisher","first-page":"298","DOI":"10.1109\/TCDS.2022.3157772","volume":"15","author":"N Sun","year":"2023","unstructured":"Sun, N., Tao, J., Liu, J., Sun, H., Han, G.: 3-D facial feature reconstruction and learning network for facial expression recognition in the wild. IEEE T. Cogn. Dev. Syst. 15(1), 298\u2013309 (2023). https:\/\/doi.org\/10.1109\/TCDS.2022.3157772","journal-title":"IEEE T. Cogn. Dev. Syst."},{"key":"4190_CR14","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.ins.2021.08.043","volume":"580","author":"Q Huang","year":"2021","unstructured":"Huang, Q., Huang, C., Wang, X., Jiang, F.: Facial expression recognition with grid-wise attention and visual transformer. Inform. Sciences. 580, 35\u201354 (2021). https:\/\/doi.org\/10.1016\/j.ins.2021.08.043","journal-title":"Inform. Sciences."},{"issue":"2","key":"4190_CR15","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1109\/MMUL.2021.3065985","volume":"28","author":"T Rao","year":"2021","unstructured":"Rao, T., Li, J., Wang, X., Sun, Y., Chen, H.: Facial expression recognition with multiscale graph convolutional networks. IEEE Multimedia 28(2), 11\u201319 (2021). https:\/\/doi.org\/10.1109\/MMUL.2021.3065985","journal-title":"IEEE Multimedia"},{"key":"4190_CR16","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., Meng, D., Qiao, Y.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 29, 4057\u20134069 (2020). https:\/\/doi.org\/10.1109\/TIP.2019.2956143","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"4190_CR17","doi-asserted-by":"publisher","first-page":"1947","DOI":"10.1177\/03611981241258753","volume":"2679","author":"X Dong","year":"2025","unstructured":"Dong, X., Shi, P., Liang, T., Yang, A.: CTAFFNet: CNN-Transformer adaptive feature fusion object detection for algorithm complex traffic scenarios. Transport. Res. Rec. 2679(1), 1947\u20131965 (2025). https:\/\/doi.org\/10.1177\/03611981241258753","journal-title":"Transport. Res. Rec."},{"key":"4190_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.displa.2024.102814","volume":"84","author":"X Dong","year":"2024","unstructured":"Dong, X., Shi, P., Qi, H., Yang, A., Liang, T.: TS-BEV: BEV object detection algorithm based on temporal-spatial feature fusion. Displays 84, 102814 (2024). https:\/\/doi.org\/10.1016\/j.displa.2024.102814","journal-title":"Displays"},{"key":"4190_CR19","doi-asserted-by":"publisher","first-page":"1640","DOI":"10.1109\/TIFS.2019.2946938","volume":"15","author":"Q Wang","year":"2020","unstructured":"Wang, Q., Guo, G.: LS-CNN: Characterizing local patches at multiple scales for face recognition. IEEE T. Inf. Foren. Sec. 15, 1640\u20131653 (2020). https:\/\/doi.org\/10.1109\/TIFS.2019.2946938","journal-title":"IEEE T. Inf. Foren. Sec."},{"issue":"2","key":"4190_CR20","doi-asserted-by":"publisher","first-page":"487","DOI":"10.1109\/JAS.2023.124029","volume":"11","author":"N Zeng","year":"2024","unstructured":"Zeng, N., Li, X., Wu, P., Li, H., Luo, X.: A novel tensor decomposition-based efficient detector for low-altitude aerial objects with knowledge distillation scheme. IEEE\/CAA J. Autom. Sin. 11(2), 487\u2013501 (2024). https:\/\/doi.org\/10.1109\/JAS.2023.124029","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"4190_CR21","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1109\/JBHI.2020.2986926","volume":"25","author":"A Sinha","year":"2021","unstructured":"Sinha, A., Dolz, J.: Multi-scale self-guided attention for medical image segmentation. IEEE J Biomed Health. 25, 121\u2013130 (2021). https:\/\/doi.org\/10.1109\/JBHI.2020.2986926","journal-title":"IEEE J Biomed Health."},{"key":"4190_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110451","volume":"269","author":"AR Shahid","year":"2023","unstructured":"Shahid, A.R., Yan, H.: SqueezExpNet: dual-stage convolutional neural network for accurate facial expression recognition with attention mechanism. Knowl-based. Syst. 269, 110451 (2023). https:\/\/doi.org\/10.1016\/j.knosys.2023.110451","journal-title":"Knowl-based. Syst."},{"issue":"3","key":"4190_CR23","doi-asserted-by":"publisher","first-page":"2918","DOI":"10.1007\/s10489-021-02575-0","volume":"52","author":"W Zou","year":"2022","unstructured":"Zou, W., Zhang, D., Lee, D.J.: A new multi-feature fusion based convolutional neural network for facial expression recognition. Appl. Intell. 52(3), 2918\u20132929 (2022). https:\/\/doi.org\/10.1007\/s10489-021-02575-0","journal-title":"Appl. Intell."},{"key":"4190_CR24","doi-asserted-by":"publisher","first-page":"410","DOI":"10.1016\/j.ins.2022.06.087","volume":"608","author":"Y Li","year":"2022","unstructured":"Li, Y., Lu, Y., Gong, M., Liu, L., Zhao, L.: Dual-channel feature disentanglement for identity-invariant facial expression recognition. Inform. Sciences. 608, 410\u2013423 (2022). https:\/\/doi.org\/10.1016\/j.ins.2022.06.087","journal-title":"Inform. Sciences."},{"key":"4190_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2023.110312","volume":"141","author":"J Xiao","year":"2023","unstructured":"Xiao, J., Gan, C., Zhu, Q., Zhu, Y., Liu, G.: CFNet: facial expression recognition via constraint fusion under multi-task joint learning network. Appl. Soft Comput. 141, 110312 (2023). https:\/\/doi.org\/10.1016\/j.asoc.2023.110312","journal-title":"Appl. Soft Comput."},{"key":"4190_CR26","doi-asserted-by":"publisher","first-page":"93","DOI":"10.1016\/j.cviu.2015.07.007","volume":"140","author":"L Zhang","year":"2015","unstructured":"Zhang, L., Mistry, K., Jiang, S., Neoh, S.C., Hossain, M.A.: Adaptive facial point detection and emotion recognition for a humanoid robot. Comput. Vis. Image Understand. 140, 93\u2013114 (2015). https:\/\/doi.org\/10.1016\/j.cviu.2015.07.007","journal-title":"Comput. Vis. Image Understand."},{"key":"4190_CR27","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1016\/j.neunet.2022.11.025","volume":"158","author":"H Gao","year":"2023","unstructured":"Gao, H., Wu, M., Chen, Z., Li, Y., Wang, X., An, S., Li, J., Liu, C.: Ssa-icl: multi-domain adaptive attention with intra-dataset continual learning for facial expression recognition. Neural Netw. 158, 228\u2013238 (2023). https:\/\/doi.org\/10.1016\/j.neunet.2022.11.025","journal-title":"Neural Netw."},{"key":"4190_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.108136","volume":"240","author":"T Ma","year":"2022","unstructured":"Ma, T., Tian, W., Xie, Y.: Multi-level knowledge distillation for low-resolution object detection and facial expression recognition. Knowl-based. Syst. 240, 108136 (2022). https:\/\/doi.org\/10.1016\/j.knosys.2022.108136","journal-title":"Knowl-based. Syst."},{"issue":"5","key":"4190_CR29","doi-asserted-by":"publisher","first-page":"3069","DOI":"10.1007\/s00530-022-00907-9","volume":"29","author":"C Fan","year":"2023","unstructured":"Fan, C., Wang, Z., Li, J., Wang, S., Sun, X.: Robust facial expression recognition with global-local joint representation learning. Multimedia Syst. 29(5), 3069\u20133079 (2023). https:\/\/doi.org\/10.1007\/s00530-022-00907-9","journal-title":"Multimedia Syst."},{"issue":"11","key":"4190_CR30","doi-asserted-by":"publisher","first-page":"2417","DOI":"10.1049\/iet-ipr.2020.0063","volume":"14","author":"S Cao","year":"2020","unstructured":"Cao, S., Yao, Y., An, G.: E2-capsule neural networks for facial expression recognition using AU-aware attention. IET Image Process. 14(11), 2417\u20132424 (2020). https:\/\/doi.org\/10.1049\/iet-ipr.2020.0063","journal-title":"IET Image Process."},{"issue":"3","key":"4190_CR31","doi-asserted-by":"publisher","first-page":"833","DOI":"10.3390\/s21030833","volume":"21","author":"X Liang","year":"2021","unstructured":"Liang, X., Xu, L., Liu, J., et al.: Patch attention layer of embedding handcrafted features in CNN for facial expression recognition. Sensors. 21(3), 833 (2021). https:\/\/doi.org\/10.3390\/s21030833","journal-title":"Sensors."},{"key":"4190_CR32","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1109\/TPAMI.2019.2938758","volume":"43","author":"S Gao","year":"2021","unstructured":"Gao, S., Cheng, M., Zhao, K., Zhang, X., Yang, M., Torr, P.: Res2net: a new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 43, 652\u2013662 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2019.2938758","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4190_CR33","doi-asserted-by":"publisher","first-page":"6544","DOI":"10.1109\/TIP.2021.3093397","volume":"30","author":"Z Zhao","year":"2021","unstructured":"Zhao, Z., Liu, Q., Wang, S.: Learning deep global multi-scale and local attention features for facial expression recognition in the wild. IEEE Trans. Image Process. 30, 6544\u20136556 (2021). https:\/\/doi.org\/10.1109\/TIP.2021.3093397","journal-title":"IEEE Trans. Image Process."},{"key":"4190_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.129323","volume":"622","author":"Z Fei","year":"2025","unstructured":"Fei, Z., Zhang, B., Zhou, W., Li, X., Zhang, Y., Fei, M.: Global multi-scale extraction and local mixed multi-head attention for facial expression recognition in the wild. Neurocomputing 622, 129323 (2025). https:\/\/doi.org\/10.1016\/j.neucom.2024.129323","journal-title":"Neurocomputing"},{"key":"4190_CR35","doi-asserted-by":"publisher","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: ECA-Net: efficient channel attention for deep convolutional neural networks. In: Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13\u201319 (2020).https:\/\/doi.org\/10.1109\/CVPR42600.2020.01155","DOI":"10.1109\/CVPR42600.2020.01155"},{"issue":"4","key":"4190_CR36","doi-asserted-by":"publisher","first-page":"5473","DOI":"10.1007\/s11042-022-12321-4","volume":"82","author":"Z He","year":"2023","unstructured":"He, Z., Meng, B., Wang, L., Jeon, G., Liu, Z., Yang, X.: Global and local fusion ensemble network for facial expression recognition. Multimed. Tools Appl. 82(4), 5473\u20135494 (2023). https:\/\/doi.org\/10.1007\/s11042-022-12321-4","journal-title":"Multimed. Tools Appl."},{"key":"4190_CR37","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-022-12321-4","volume":"139","author":"Y Liu","year":"2023","unstructured":"Liu, Y., Peng, J., Dai, W., Zeng, J., Shan, S.: Joint spatial and scale attention network for multi-view facial expression recognition. Pattern Recognit. 139, 109496 (2023). https:\/\/doi.org\/10.1007\/s11042-022-12321-4","journal-title":"Pattern Recognit."},{"key":"4190_CR38","doi-asserted-by":"publisher","first-page":"179656","DOI":"10.1109\/ACCESS.2020.3025372","volume":"8","author":"T Fan","year":"2020","unstructured":"Fan, T., Wang, G., Li, Y., Wang, H.: Ma-net: a multi-scale attention network for liver and tumor segmentation. IEEE Access. 8, 179656\u2013179665 (2020). https:\/\/doi.org\/10.1109\/ACCESS.2020.3025372","journal-title":"IEEE Access."},{"key":"4190_CR39","doi-asserted-by":"publisher","unstructured":"Woo, S., Papk, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018). https:\/\/doi.org\/10.48550\/arXiv.1807.06521","DOI":"10.48550\/arXiv.1807.06521"},{"key":"4190_CR40","doi-asserted-by":"publisher","first-page":"1489","DOI":"10.1109\/TPAMI.2022.3164083","volume":"45","author":"Y Li","year":"2022","unstructured":"Li, Y., Yao, T., Pan, Y., Mei, T.: Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 45, 1489\u20131500 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3164083","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4190_CR41","doi-asserted-by":"publisher","first-page":"21625","DOI":"10.1007\/s00521-021-06778-x","volume":"34","author":"N Le","year":"2022","unstructured":"Le, N., Nguyen, K., Nguyen, A., Le, B.: Global-local attention for emotion recognition. Neural Comput. Appl. 34, 21625\u201321639 (2022). https:\/\/doi.org\/10.1007\/s00521-021-06778-x","journal-title":"Neural Comput. Appl."},{"key":"4190_CR42","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2023.119301","volume":"644","author":"X Xia","year":"2023","unstructured":"Xia, X., Jiang, D.: HiT-MST: Dynamic facial expression recognition with hierarchical transformers and multi-scale spatiotemporal aggregation. Inf. Sci. 644, 119301 (2023). https:\/\/doi.org\/10.1016\/j.ins.2023.119301","journal-title":"Inf. Sci."},{"key":"4190_CR43","doi-asserted-by":"publisher","first-page":"781","DOI":"10.1016\/j.ins.2022.11.068","volume":"619","author":"C Liu","year":"2023","unstructured":"Liu, C., Hirota, K., Dai, Y.: Patch attention convolutional vision transformer for facial expression recognition with occlusion. Inf. Sci. 619, 781\u2013794 (2023). https:\/\/doi.org\/10.1016\/j.ins.2022.11.068","journal-title":"Inf. Sci."},{"key":"4190_CR44","doi-asserted-by":"publisher","unstructured":"Huang, Z., Zhang, D., Ren, F., Hu, M. Liu, J., Yu, H.: SG-TE: Spatial guidance and temporal enhancement network for facial-bodily emotion recognition. CAAI T. Intell. Techno. pp,1\u201320 (2025). https:\/\/doi.org\/10.1049\/cit2.70006.","DOI":"10.1049\/cit2.70006"},{"key":"4190_CR45","doi-asserted-by":"publisher","unstructured":"Sarwo: Improving offline handwritten text recognition accuracy with ADAM and SGD optimizers and convolutional neural networks models. In: 2024 International Conference on Intelligent Cybernetics Technology & Applications (ICICyTA),pp. 1135\u20131139 (2024).https:\/\/doi.org\/10.1109\/ICICYTA64807.2024.10913452","DOI":"10.1109\/ICICYTA64807.2024.10913452"},{"key":"4190_CR46","doi-asserted-by":"publisher","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition(CVPR), pp. 21\u201326 (2017). https:\/\/doi.org\/10.1109\/tip.2018.2868382","DOI":"10.1109\/tip.2018.2868382"},{"key":"4190_CR47","doi-asserted-by":"publisher","unstructured":"Barsoum, E., Zhang, C., Ferrer, C.C., Zhang, Z.: Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction. Association for Computing Machinery. pp. 279\u2013283 (2016). https:\/\/doi.org\/10.1145\/2993148.2993165","DOI":"10.1145\/2993148.2993165"},{"key":"4190_CR48","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition(CVPR). pp. 27\u201330 (2016). https:\/\/doi.org\/10.48550\/arXiv.1512.03385","DOI":"10.48550\/arXiv.1512.03385"},{"key":"4190_CR49","doi-asserted-by":"publisher","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), pp. 22\u201329 (2017). https:\/\/doi.org\/10.1007\/s11263-019-01228-7","DOI":"10.1007\/s11263-019-01228-7"},{"key":"4190_CR50","doi-asserted-by":"publisher","unstructured":"Khan, R. A., Meyer, A., Konik, H., Bouakaz, S.: Exploring human visual system: study to aid the development of automatic facial expression recognition framework. In: Proceedings of 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 49\u201354 (2012). https:\/\/doi.org\/10.1109\/CVPRW.2012.6239186","DOI":"10.1109\/CVPRW.2012.6239186"},{"key":"4190_CR51","doi-asserted-by":"publisher","first-page":"159","DOI":"10.1016\/j.neucom.2020.12.076","volume":"432","author":"H Li","year":"2021","unstructured":"Li, H., Wang, N., Yu, Y., Yang, X., Gao, X.: LBAN-IL: a novel method of high discriminative representation for facial expression recognition. Neurocomputing 432, 159\u2013169 (2021). https:\/\/doi.org\/10.1016\/j.neucom.2020.12.076","journal-title":"Neurocomputing"},{"issue":"2","key":"4190_CR52","doi-asserted-by":"publisher","first-page":"1236","DOI":"10.1109\/TAFFC.2021.3122146","volume":"14","author":"F Ma","year":"2023","unstructured":"Ma, F., Sun, B., Li, S.: Facial expression recognition with visual transformers and attentional selective fusion. IEEE T. Affect. Comput. 14(2), 1236\u20131248 (2023). https:\/\/doi.org\/10.1109\/TAFFC.2021.3122146","journal-title":"IEEE T. Affect. Comput."},{"key":"4190_CR53","doi-asserted-by":"publisher","unstructured":"Zhang, X., Lu, Y., Yan, H. Huang, J., Gu, Y., Ji, Y., Liu, Z, Liu, B.: ReSup: reliable label noise suppression for facial expression recognition. IEEE T. Affect. Comput. pp.1\u201314(2025). https:\/\/doi.org\/10.1109\/TAFFC.2025.3549017","DOI":"10.1109\/TAFFC.2025.3549017"},{"issue":"12","key":"4190_CR54","doi-asserted-by":"publisher","first-page":"8693","DOI":"10.1007\/s11760-024-03501-w","volume":"18","author":"W Dong","year":"2024","unstructured":"Dong, W., Zheng, X., Zhang, L., Zhang, Y.: Attentional visual graph neural network based facial expression recognition method. Signal Image Video Process. 18(12), 8693\u20138705 (2024). https:\/\/doi.org\/10.1007\/s11760-024-03501-w","journal-title":"Signal Image Video Process."},{"key":"4190_CR55","doi-asserted-by":"publisher","unstructured":"Zhang, H., Zhuang, X., Gao, X., Mao, R., Ren, Q.: Lightweight facial expression recognition based on hybrid multiscale and multi-head collaborative attention. In: Chinese Conference on Pattern Recognition and Computer Vision (PRCV), pp. 304\u2013316 (2025). https:\/\/doi.org\/10.1007\/978-981-97-8490-5_22","DOI":"10.1007\/978-981-97-8490-5_22"},{"key":"4190_CR56","doi-asserted-by":"publisher","unstructured":"Ding, H., Zhou, P., Chellappa, R.: Occlusion-adaptive deep network for robust facial expression recognition. In: Proceedings of the 2020 IEEE International Joint Conference on Biometrics (IJCB), pp. 1\u20139 (2020). https:\/\/doi.org\/10.1109\/IJCB48548.2020.9304923","DOI":"10.1109\/IJCB48548.2020.9304923"},{"key":"4190_CR57","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2024.104062","volume":"98","author":"Y Tan","year":"2024","unstructured":"Tan, Y., Xia, H., Song, S.: Learning informative and discriminative semantic features for robust facial expression recognition. J. Vis. Commun. Image Represent. 98, 104062 (2024). https:\/\/doi.org\/10.1016\/j.jvcir.2024.104062","journal-title":"J. Vis. Commun. Image Represent."},{"key":"4190_CR58","doi-asserted-by":"publisher","unstructured":"Liu, W., Shi, X., Liu, X.: Flipping consistent and counterfactual attention network for facial expression recognition. In: Proceedings of the 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2665\u20132669 (2024). https:\/\/doi.org\/10.1109\/ICASSP48485.2024.10448412","DOI":"10.1109\/ICASSP48485.2024.10448412"},{"issue":"1","key":"4190_CR59","doi-asserted-by":"publisher","DOI":"10.1007\/s00138-024-01641-0","volume":"36","author":"G Devasena","year":"2025","unstructured":"Devasena, G., Vidhya, V.: Twinned attention network for occlusion-aware facial expression recognition. Mach. Vis. Appl. 36(1), 23 (2025). https:\/\/doi.org\/10.1007\/s00138-024-01641-0","journal-title":"Mach. Vis. Appl."},{"key":"4190_CR60","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2019","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: Affectnet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10, 18\u201331 (2019). https:\/\/doi.org\/10.1109\/TAFFC.2017.2740923","journal-title":"IEEE Trans. Affect. Comput."},{"key":"4190_CR61","doi-asserted-by":"publisher","unstructured":"Wang, X., Xie, L., Dong, C., Shan, Y.: Real-ESRGAN: Training real-world blind super-resolution with pure synthetic data. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 1905\u20131914(2021). https:\/\/doi.org\/10.1109\/ICCVW54120.2021.00217","DOI":"10.1109\/ICCVW54120.2021.00217"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04190-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04190-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04190-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:16:00Z","timestamp":1763644560000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04190-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,24]]},"references-count":61,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4190"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04190-3","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2025,9,24]]},"assertion":[{"value":"8 October 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"These authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence their work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}