{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,9]],"date-time":"2026-05-09T07:26:40Z","timestamp":1778311600528,"version":"3.51.4"},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,6,10]],"date-time":"2023-06-10T00:00:00Z","timestamp":1686355200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,6,10]],"date-time":"2023-06-10T00:00:00Z","timestamp":1686355200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62106054"],"award-info":[{"award-number":["62106054"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Science and Technology Project of Guangxi","award":["2018GXNSFAA281351"],"award-info":[{"award-number":["2018GXNSFAA281351"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,3]]},"DOI":"10.1007\/s00371-023-02900-3","type":"journal-article","created":{"date-parts":[[2023,6,10]],"date-time":"2023-06-10T09:02:22Z","timestamp":1686387742000},"page":"2035-2047","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":20,"title":["Feature fusion of multi-granularity and multi-scale for facial expression recognition"],"prefix":"10.1007","volume":"40","author":[{"given":"Haiying","family":"Xia","sequence":"first","affiliation":[]},{"given":"Lidan","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Shuxiang","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,10]]},"reference":[{"issue":"3","key":"2900_CR1","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li, S., Deng, W.: Deep facial expression recognition: a survey. IEEE Trans. Affect. Comput. 13(3), 1195\u20131215 (2020). https:\/\/doi.org\/10.1109\/TAFFC.2020.2981446","journal-title":"IEEE Trans. Affect. Comput."},{"key":"2900_CR2","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., et\u00a0al.: The extended cohn-kanade dataset (ck+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops, pp. 94\u2013101 (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"issue":"9","key":"2900_CR3","doi-asserted-by":"publisher","first-page":"607","DOI":"10.1016\/j.imavis.2011.07.002","volume":"29","author":"G Zhao","year":"2011","unstructured":"Zhao, G., Huang, X., Taini, M., Li, S.Z., Pietik\u00e4Inen, M.: Facial expression recognition from near-infrared videos. Image Vis. Comput. 29(9), 607\u2013619 (2011). https:\/\/doi.org\/10.1016\/j.imavis.2011.07.002","journal-title":"Image Vis. Comput."},{"key":"2900_CR4","unstructured":"Pantic, M., Valstar, M., Rademaker, R., Maat, L.: Web-based database for facial expression analysis. In: 2005 IEEE International Conference on Multimedia and Expo, p. 5 (2005)"},{"key":"2900_CR5","unstructured":"Kim, Y., Yoo, B., Kwak, Y., Choi, C., Kim, J.: Deep generative-contrastive networks for facial expression recognition. arXiv preprint (2017). arXiv:1703.07140"},{"issue":"9","key":"2900_CR6","doi-asserted-by":"publisher","first-page":"4193","DOI":"10.1109\/TIP.2017.2689999","volume":"26","author":"K Zhang","year":"2017","unstructured":"Zhang, K., Huang, Y., Du, Y., Wang, L.: Facial expression recognition based on deep evolutional spatial-temporal networks. IEEE Trans. Image Process. 26(9), 4193\u20134203 (2017). https:\/\/doi.org\/10.1109\/TIP.2017.2689999","journal-title":"IEEE Trans. Image Process."},{"key":"2900_CR7","doi-asserted-by":"crossref","unstructured":"Yang, H., Ciftci, U., Yin, L.: Facial expression recognition by de-expression residue learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2168\u20132177 (2018)","DOI":"10.1109\/CVPR.2018.00231"},{"key":"2900_CR8","doi-asserted-by":"crossref","unstructured":"Hazourli, A.R., Djeghri, A., Salam, H., Othmani, A.: Deep multi-facial patches aggregation network for facial expression recognition. arXiv preprint (2020). arXiv:2002.09298","DOI":"10.1007\/s11042-020-10332-7"},{"key":"2900_CR9","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"2900_CR10","doi-asserted-by":"crossref","unstructured":"Goodfellow, I.J., Erhan, D., Carrier, P.L., et\u00a0al.: Challenges in representation learning: A report on three machine learning contests. In: International Conference on Neural Information Processing, pp. 117\u2013124 (2013)","DOI":"10.1007\/978-3-642-42051-1_16"},{"issue":"1","key":"2900_CR11","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2019","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: Affectnet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18\u201331 (2019). https:\/\/doi.org\/10.1109\/TAFFC.2017.2740923","journal-title":"IEEE Trans. Affect. Comput."},{"key":"2900_CR12","doi-asserted-by":"crossref","unstructured":"Farzaneh, A.H., Qi, X.: Facial expression recognition in the wild via deep attentive center loss. In: 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 2401\u20132410 (2021)","DOI":"10.1109\/WACV48630.2021.00245"},{"key":"2900_CR13","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-022-02619-7","author":"H Li","year":"2022","unstructured":"Li, H., Xiao, X., Liu, X., Guo, J., Wen, G., Liang, P.: Heuristic objective for facial expression recognition. Vis. Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-022-02619-7","journal-title":"Vis. Comput."},{"key":"2900_CR14","doi-asserted-by":"crossref","unstructured":"Siqueira, H., Magg, S., Wermter, S.: Efficient facial feature learning with wide ensemble-based convolutional neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 5800\u20135809 (2020)","DOI":"10.1609\/aaai.v34i04.6037"},{"key":"2900_CR15","doi-asserted-by":"crossref","unstructured":"Cai, J., Meng, Z., Khan, A.S., et\u00a0al.: Identity-free facial expression recognition using conditional generative adversarial network. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 1344\u20131348 (2021)","DOI":"10.1109\/ICIP42928.2021.9506593"},{"key":"2900_CR16","doi-asserted-by":"crossref","unstructured":"Zhang, F., Zhang, T., Mao, Q., Xu, C.: Joint pose and expression modeling for facial expression recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3359\u20133368 (2018)","DOI":"10.1109\/CVPR.2018.00354"},{"issue":"2","key":"2900_CR17","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1167\/9.2.23","volume":"9","author":"Z Hammal","year":"2009","unstructured":"Hammal, Z., Arguin, M., Gosselin, F.: Comparing a novel model based on the transferable belief model with humans during the recognition of partially occluded facial expressions. J. Vis. 9(2), 22\u201322 (2009). https:\/\/doi.org\/10.1167\/9.2.23","journal-title":"J. Vis."},{"key":"2900_CR18","doi-asserted-by":"crossref","unstructured":"Ram\u00edrez\u00a0Cornejo, J.Y., Pedrini, H.: Recognition of occluded facial expressions based on centrist features. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1298\u20131302 (2016)","DOI":"10.1109\/ICASSP.2016.7471886"},{"key":"2900_CR19","doi-asserted-by":"crossref","unstructured":"Pan, B., Wang, S., Xia, B.: Occluded facial expression recognition enhanced through privileged information. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 566\u2013573 (2019)","DOI":"10.1145\/3343031.3351049"},{"key":"2900_CR20","doi-asserted-by":"crossref","unstructured":"Adil, B., Nadjib, K.M., Yacine, L.: A novel approach for facial expression recognition. In: 2019 International Conference on Networking and Advanced Systems (ICNAS), pp. 1\u20135 (2019)","DOI":"10.1109\/ICNAS.2019.8807883"},{"key":"2900_CR21","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Liu, Q., Zhou, F.: Robust lightweight facial expression recognition network with label distribution training. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 3510\u20133519 (2021)","DOI":"10.1609\/aaai.v35i4.16465"},{"key":"2900_CR22","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., Meng, D., Qiao, Y.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 29, 4057\u20134069 (2020). https:\/\/doi.org\/10.1109\/TIP.2019.2956143","journal-title":"IEEE Trans. Image Process."},{"issue":"5","key":"2900_CR23","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/TIP.2018.2886767","volume":"28","author":"Y Li","year":"2019","unstructured":"Li, Y., Zeng, J., Shan, S., Chen, X.: Occlusion aware facial expression recognition using cnn with attention mechanism. IEEE Trans. Image Process. 28(5), 2439\u20132450 (2019). https:\/\/doi.org\/10.1109\/TIP.2018.2886767","journal-title":"IEEE Trans. Image Process."},{"key":"2900_CR24","doi-asserted-by":"crossref","unstructured":"Du, R., Chang, D., Bhunia, A.K., Xie, J., Ma, Z., Song, Y.-Z., Guo, J.: Fine-grained visual classification via progressive multi-granularity training of jigsaw patches. In: European Conference on Computer Vision, pp. 153\u2013168 (2020)","DOI":"10.1007\/978-3-030-58565-5_10"},{"key":"2900_CR25","doi-asserted-by":"crossref","unstructured":"Ding, H., Zhou, P., Chellappa, R.: Occlusion-adaptive deep network for robust facial expression recognition. In: 2020 IEEE International Joint Conference on Biometrics (IJCB), pp. 1\u20139 (2020)","DOI":"10.1109\/IJCB48548.2020.9304923"},{"key":"2900_CR26","doi-asserted-by":"publisher","first-page":"6544","DOI":"10.1109\/TIP.2021.3093397","volume":"30","author":"Z Zhao","year":"2021","unstructured":"Zhao, Z., Liu, Q., Wang, S.: Learning deep global multi-scale and local attention features for facial expression recognition in the wild. IEEE Trans. Image Process. 30, 6544\u20136556 (2021). https:\/\/doi.org\/10.1109\/TIP.2021.3093397","journal-title":"IEEE Trans. Image Process."},{"key":"2900_CR27","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2021.3122146","author":"F Ma","year":"2021","unstructured":"Ma, F., Sun, B., Li, S.: Facial expression recognition with visual transformers and attentional selective fusion. IEEE Trans. Affect. Comput. (2021). https:\/\/doi.org\/10.1109\/TAFFC.2021.3122146","journal-title":"IEEE Trans. Affect. Comput."},{"key":"2900_CR28","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-022-02413-5","author":"X Liang","year":"2022","unstructured":"Liang, X., Xu, L., Zhang, W., et al.: A convolution-transformer dual branch network for head-pose and occlusion facial expression recognition. Vis. Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-022-02413-5","journal-title":"Vis. Comput."},{"key":"2900_CR29","doi-asserted-by":"publisher","first-page":"781","DOI":"10.1016\/j.ins.2022.11.068","volume":"619","author":"C Liu","year":"2023","unstructured":"Liu, C., Hirota, K., Dai, Y.: Patch attention convolutional vision transformer for facial expression recognition with occlusion. Inf. Sci. 619, 781\u2013794 (2023). https:\/\/doi.org\/10.1016\/j.ins.2022.11.068","journal-title":"Inf. Sci."},{"issue":"3","key":"2900_CR30","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1007\/s00138-022-01288-9","volume":"33","author":"L Liao","year":"2022","unstructured":"Liao, L., Zhu, Y., Zheng, B., Jiang, X., Lin, J.: Fergcn: facial expression recognition based on graph convolution network. Mach. Vis. Appl. 33(3), 40 (2022). https:\/\/doi.org\/10.1007\/s00138-022-01288-9","journal-title":"Mach. Vis. Appl."},{"key":"2900_CR31","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1016\/j.neunet.2022.11.025","volume":"158","author":"H Gao","year":"2023","unstructured":"Gao, H., Wu, M., Chen, Z., et al.: Ssa-icl: Multi-domain adaptive attention with intra-dataset continual learning for facial expression recognition. Neural Netw. 158, 228\u2013238 (2023). https:\/\/doi.org\/10.1016\/j.neunet.2022.11.025","journal-title":"Neural Netw."},{"key":"2900_CR32","doi-asserted-by":"crossref","unstructured":"Ruan, D., Yan, Y., Lai, S., et\u00a0al.: Feature decomposition and reconstruction learning for effective facial expression recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7656\u20137665 (2021)","DOI":"10.1109\/CVPR46437.2021.00757"},{"key":"2900_CR33","doi-asserted-by":"crossref","unstructured":"Wang, K., Peng, X., Yang, J., Lu, S., Qiao, Y.: Suppressing uncertainties for large-scale facial expression recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6897\u20136906 (2020)","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"2900_CR34","doi-asserted-by":"crossref","unstructured":"Noroozi, M., Favaro, P.: Unsupervised learning of visual representations by solving jigsaw puzzles. In: European Conference on Computer Vision, pp. 69\u201384 (2016)","DOI":"10.1007\/978-3-319-46466-4_5"},{"key":"2900_CR35","doi-asserted-by":"crossref","unstructured":"Chen, Y., Bai, Y., Zhang, W., Mei, T.: Destruction and construction learning for fine-grained image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5152\u20135161 (2019)","DOI":"10.1109\/CVPR.2019.00530"},{"issue":"2","key":"2900_CR36","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1109\/MMUL.2021.3076834","volume":"28","author":"H Xia","year":"2021","unstructured":"Xia, H., Li, C., Tan, Y., Li, L., Song, S.: Destruction and reconstruction learning for facial expression recognition. IEEE Multimed. 28(2), 20\u201328 (2021). https:\/\/doi.org\/10.1109\/MMUL.2021.3076834","journal-title":"IEEE Multimed."},{"issue":"6","key":"2900_CR37","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017). https:\/\/doi.org\/10.1145\/3065386","journal-title":"Commun. ACM"},{"key":"2900_CR38","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint (2014). arXiv:1409.1556"},{"key":"2900_CR39","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., et\u00a0al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"2900_CR40","unstructured":"Duta, I.C., Liu, L., Zhu, F., Shao, L.: Pyramidal convolution: rethinking convolutional neural networks for visual recognition. arXiv preprint (2020). arXiv:2006.11538"},{"issue":"2","key":"2900_CR41","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1109\/TPAMI.2019.2938758","volume":"43","author":"S Gao","year":"2021","unstructured":"Gao, S., Cheng, M., Zhao, K., et al.: Res2net: a new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 43(2), 652\u2013662 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2019.2938758","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2900_CR42","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"2900_CR43","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"2900_CR44","doi-asserted-by":"crossref","unstructured":"Barsoum, E., Zhang, C., Ferrer, C.C., Zhang, Z.: Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction, pp. 279\u2013283 (2016)","DOI":"10.1145\/2993148.2993165"},{"key":"2900_CR45","doi-asserted-by":"crossref","unstructured":"Guo, Y., Zhang, L., Hu, Y., He, X., Gao, J.: Ms-celeb-1m: a dataset and benchmark for large-scale face recognition. In: European Conference on Computer Vision, pp. 87\u2013102 (2016)","DOI":"10.1007\/978-3-319-46487-9_6"},{"key":"2900_CR46","unstructured":"Paszke, A., Gross, S., Massa, F., et\u00a0al.: Pytorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst., vol. 32 (2019)"},{"key":"2900_CR47","doi-asserted-by":"crossref","unstructured":"Huang, C.: Combining convolutional neural networks for emotion recognition. In: 2017 IEEE MIT Undergraduate Research Technology Conference (URTC), pp. 1\u20134 (2017)","DOI":"10.1109\/URTC.2017.8284175"},{"key":"2900_CR48","doi-asserted-by":"publisher","DOI":"10.1007\/s10044-022-01124-w","author":"C Su","year":"2022","unstructured":"Su, C., Wei, J., Lin, D., Kong, L.: Using attention lsgb network for facial expression recognition. Pattern Anal. Appl. (2022). https:\/\/doi.org\/10.1007\/s10044-022-01124-w","journal-title":"Pattern Anal. Appl."},{"key":"2900_CR49","doi-asserted-by":"crossref","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., Balasubramanian, V.N.: Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 839\u2013847 (2018)","DOI":"10.1109\/WACV.2018.00097"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02900-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-02900-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02900-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,18]],"date-time":"2024-02-18T23:34:09Z","timestamp":1708299249000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-02900-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,10]]},"references-count":49,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,3]]}},"alternative-id":["2900"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-02900-3","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,10]]},"assertion":[{"value":"10 May 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 June 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}