{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T22:50:45Z","timestamp":1768776645551,"version":"3.49.0"},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T00:00:00Z","timestamp":1735862400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T00:00:00Z","timestamp":1735862400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"The Key Scientific Research Projects Applied Research Plan in Henan Colleges and Universities","award":["22A510013"],"award-info":[{"award-number":["22A510013"]}]},{"name":"The Open Subject of Scientific Research Platform in Grain Information Processing Center","award":["KFJJ2022011"],"award-info":[{"award-number":["KFJJ2022011"]}]},{"name":"The Innovative Funds Plan of Henan University of Technology","award":["2022ZKCJ13"],"award-info":[{"award-number":["2022ZKCJ13"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s00530-024-01628-x","type":"journal-article","created":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T15:33:18Z","timestamp":1735918398000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Facial expression recognition via joint loss constraining attention-modulated contextual spatial information network"],"prefix":"10.1007","volume":"31","author":[{"given":"Xue","family":"Li","sequence":"first","affiliation":[]},{"given":"Chunhua","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Huawei","family":"Tao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,3]]},"reference":[{"key":"1628_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.cmpb.2022.106621","volume":"215","author":"H Ge","year":"2022","unstructured":"Ge, H., Zhu, Z., Dai, Y., et al.: Facial expression recognition based on deep learning. Comput. Methods Prog. Biomed. 215, 1\u20139 (2022). https:\/\/doi.org\/10.1016\/j.cmpb.2022.106621","journal-title":"Comput. Methods Prog. Biomed."},{"issue":"8","key":"1628_CR2","doi-asserted-by":"publisher","first-page":"5619","DOI":"10.1109\/TII.2022.3141400","volume":"18","author":"C Bisogni","year":"2022","unstructured":"Bisogni, C., Castiglione, A., Hossain, S., et al.: Impact of deep learning approaches on facial expression recognition in healthcare industries. IEEE Trans. Ind. Inform. 18(8), 5619\u20135627 (2022). https:\/\/doi.org\/10.1109\/TII.2022.3141400","journal-title":"IEEE Trans. Ind. Inform."},{"key":"1628_CR3","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1007\/s00530-022-00984-w","volume":"29","author":"RR Adyapady","year":"2022","unstructured":"Adyapady, R.R., Annappa, B.: A comprehensive review of facial expression recognition techniques. Multimed. Syst. 29, 73\u2013103 (2022). https:\/\/doi.org\/10.1007\/s00530-022-00984-w","journal-title":"Multimed. Syst."},{"key":"1628_CR4","doi-asserted-by":"crossref","unstructured":"Bhattacharya, S.: A survey on: facial expression recognition using various deep learning techniques. In: Proceedings of the ICACCP 2021 on Advanced Computational Paradigms and Hybrid Intelligent Computing, pp. 619\u2013631 (2022)","DOI":"10.1007\/978-981-16-4369-9_59"},{"key":"1628_CR5","first-page":"0759","volume":"25","author":"SS Sun","year":"2020","unstructured":"Sun, S.S., Tian, D., Wu, W., et al.: Lung segmentation by active shape model approach based on low rank theory. Int. J. Image Graph 25, 0759\u20130767 (2020)","journal-title":"Int. J. Image Graph"},{"issue":"02","key":"1628_CR6","doi-asserted-by":"publisher","first-page":"338","DOI":"10.11999\/JEIT170416","volume":"40","author":"LF Xu","year":"2018","unstructured":"Xu, L.F., Wang, J.Y., Cui, J.N., et al.: Dynamic expression recognition based on dynamic time warping and active appearance model. J. Electron. Inf. Technol. 40(02), 338\u2013345 (2018). https:\/\/doi.org\/10.11999\/JEIT170416","journal-title":"J. Electron. Inf. Technol."},{"key":"1628_CR7","doi-asserted-by":"publisher","first-page":"1638","DOI":"10.1109\/TNNLS.2019.2921564","volume":"31","author":"S Jia","year":"2019","unstructured":"Jia, S., Lin, Z., Deng, B., et al.: Cascade superpixel regularized Gabor feature fusion for hyperspectral image classification. IEEE Trans. Neural. Netw. Learn. 31, 1638\u20131652 (2019). https:\/\/doi.org\/10.1109\/TNNLS.2019.2921564","journal-title":"IEEE Trans. Neural. Netw. Learn."},{"key":"1628_CR8","doi-asserted-by":"publisher","first-page":"2243","DOI":"10.1007\/s11042-020-09663-2","volume":"88","author":"DGR Kola","year":"2021","unstructured":"Kola, D.G.R., Samayamantula, S.K.: A novel approach for facial expression recognition using local binary pattern with adaptive window. Multimed. Tools Appl. 88, 2243\u20132262 (2021). https:\/\/doi.org\/10.1007\/s11042-020-09663-2","journal-title":"Multimed. Tools Appl."},{"issue":"5","key":"1628_CR9","doi-asserted-by":"publisher","first-page":"1127","DOI":"10.1007\/s11390-020-9665-4","volume":"35","author":"A Caroppo","year":"2020","unstructured":"Caroppo, A., Leone, A., Siciliano, P.: Comparison between deep learning models and traditional machine learning approaches for facial expression recognition in ageing adults. J. Comput. Sci. Technol. 35(5), 1127\u20131146 (2020). https:\/\/doi.org\/10.1007\/s11390-020-9665-4","journal-title":"J. Comput. Sci. Technol."},{"key":"1628_CR10","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.neucom.2019.05.005","volume":"355","author":"J Shao","year":"2019","unstructured":"Shao, J., Qian, Y.: Three convolutional neural network models for facial expression recognition in the wild. Neurocomputing 355, 82\u201392 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.05.005","journal-title":"Neurocomputing"},{"key":"1628_CR11","doi-asserted-by":"crossref","unstructured":"Wang, K., Peng, X., Yang, J.: Suppressing uncertainties for large-scale facial expression recognition. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6897\u20136906 (2020)","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"1628_CR12","doi-asserted-by":"crossref","unstructured":"Zhang, W., Ji, X., Chen, K., et al.: Learning a facial expression embedding disentangled from identity. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6759\u20136768 (2021)","DOI":"10.1109\/CVPR46437.2021.00669"},{"issue":"7","key":"1628_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/e24070882","volume":"27","author":"X Li","year":"2022","unstructured":"Li, X., Zhu, C., Zhou, F.: Facial expression recognition: one attention-modulated contextual spatial information network. Entropy 27(7), 1\u201313 (2022). https:\/\/doi.org\/10.3390\/e24070882","journal-title":"Entropy"},{"key":"1628_CR14","doi-asserted-by":"crossref","unstructured":"Li, X., Zhu, C., Zhou, F.: Relation-aware facial expression recognition using contextual residual network with attention mechanism. In: Proceedings of the 17th International Conference on Computer Science and Education (ICCSE), pp. 642\u2013651 (2023)","DOI":"10.1007\/978-981-99-2443-1_55"},{"key":"1628_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.eswa.2021.116046","volume":"189","author":"Q Zhu","year":"2022","unstructured":"Zhu, Q., Mao, Q., Jia, H., et al.: Convolutional relation network for facial expression recognition in the wild with few-shot learning. Expert Syst. Appl. 189, 1\u20139 (2022). https:\/\/doi.org\/10.1016\/j.eswa.2021.116046","journal-title":"Expert Syst. Appl."},{"issue":"4","key":"1628_CR16","doi-asserted-by":"publisher","first-page":"2086","DOI":"10.1109\/TAFFC.2022.3184995","volume":"13","author":"M Jampour","year":"2022","unstructured":"Jampour, M., Javidi, M.: Multiview facial expression recognition, a survey. IEEE Trans. Affect. Comput. 13(4), 2086\u20132105 (2022). https:\/\/doi.org\/10.1109\/TAFFC.2022.3184995","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1628_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.bspc.2022.104209","volume":"79","author":"H Liu","year":"2023","unstructured":"Liu, H., Cai, H., Lin, Q., et al.: FEDA: fine-grained emotion difference analysis for facial expression recognition. Biomed Signal Proces 79, 1\u201311 (2023). https:\/\/doi.org\/10.1016\/j.bspc.2022.104209","journal-title":"Biomed Signal Proces"},{"key":"1628_CR18","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2023.3234312","author":"D Chen","year":"2023","unstructured":"Chen, D., Wen, G., Li, H., et al.: Multi-relations aware network for in-the-wild facial expression recognition. IEEE Trans. Circ. Syst. Vid. (Early Access) (2023). https:\/\/doi.org\/10.1109\/TCSVT.2023.3234312","journal-title":"IEEE Trans. Circ. Syst. Vid. (Early Access)"},{"key":"1628_CR19","doi-asserted-by":"crossref","unstructured":"Chen, J., Luo, X., Meng, Z., et al.: Research on facial expression recognition based on improved deep residual network model. In: Proceedings of the 4th International Conference on Computer Information Science and Application Technology (CISAT), pp. 1\u20137 (2021)","DOI":"10.1088\/1742-6596\/2010\/1\/012139"},{"key":"1628_CR20","doi-asserted-by":"crossref","unstructured":"Chopra, S., Hadsell, R., LeCun, Y.: Learning a similarity metric discriminatively, with application to face verification. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), pp. 539\u2013546 (2005)","DOI":"10.1109\/CVPR.2005.202"},{"key":"1628_CR21","doi-asserted-by":"crossref","unstructured":"Hadsell, R., Chopra, S., LeCun, Y.: Dimensionality reduction by learning an invariant mapping. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1735\u20131742 (2006)","DOI":"10.1109\/CVPR.2006.100"},{"key":"1628_CR22","doi-asserted-by":"crossref","unstructured":"Schroff, F., Kalenichenko, D., Philbin, J.: Facenet: a unified embedding for face recognition and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 815\u2013823 (2015)","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"1628_CR23","doi-asserted-by":"crossref","unstructured":"Wen, Y., Zhang, K., Li, Z., et al.: A discriminative feature learning approach for deep face recognition. In: Proceedings of the 14th European Conference on Computer Vision (ECCV), pp. 499\u2013515 (2016)","DOI":"10.1007\/978-3-319-46478-7_31"},{"key":"1628_CR24","doi-asserted-by":"crossref","unstructured":"Liu, W., Wen, Y., Yu, Z., et al.: Sphereface: deep hypersphere embedding for face recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 212\u2013220 (2017)","DOI":"10.1109\/CVPR.2017.713"},{"key":"1628_CR25","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Y., Zhou, Z., et al.: Cosface: large margin cosine loss for deep face recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 5265\u20135274 (2018)","DOI":"10.1109\/CVPR.2018.00552"},{"issue":"7","key":"1628_CR26","doi-asserted-by":"publisher","first-page":"926","DOI":"10.1109\/LSP.2018.2822810","volume":"25","author":"F Wang","year":"2018","unstructured":"Wang, F., Cheng, J., Liu, W., et al.: Additive margin softmax for face verification. IEEE Signal Process. Lett. 25(7), 926\u2013930 (2018). https:\/\/doi.org\/10.1109\/LSP.2018.2822810","journal-title":"IEEE Signal Process. Lett."},{"key":"1628_CR27","doi-asserted-by":"crossref","unstructured":"Deng, J., Guo, J., Xue, N., et al.: Arcface: additive angular margin loss for deep face recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4690\u20134699 (2019)","DOI":"10.1109\/CVPR.2019.00482"},{"key":"1628_CR28","doi-asserted-by":"crossref","unstructured":"Cai, J., Meng, Z., Khan, A.S., et al.: Island loss for learning discriminative features in facial expression recognition. In: Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 302\u2013309 (2018)","DOI":"10.1109\/FG.2018.00051"},{"key":"1628_CR29","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.P.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"1628_CR30","doi-asserted-by":"crossref","unstructured":"Guo, Y., Tao, D., Yu, J., et al.: Deep neural networks with relativity learning for facial expression recognition. In: Proceedings of the IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 1\u20136 (2016)","DOI":"10.1109\/ICMEW.2016.7574736"},{"key":"1628_CR31","doi-asserted-by":"crossref","unstructured":"Zhao, X., Liang, X., Liu, L., et al.: Peak-piloted deep network for facial expression recognition. In: Proceedings of the 14th European Conference on Computer Vision (ECCV), pp. 425\u2013442 (2016)","DOI":"10.1007\/978-3-319-46475-6_27"},{"key":"1628_CR32","doi-asserted-by":"publisher","first-page":"26756","DOI":"10.1109\/ACCESS.2022.3156598","volume":"10","author":"AP Fard","year":"2022","unstructured":"Fard, A.P., Mahoor, M.H.: Ad-corre: adaptive correlation-based loss for facial expression recognition in the wild. IEEE Access 10, 26756\u201326768 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2022.3156598","journal-title":"IEEE Access"},{"key":"1628_CR33","doi-asserted-by":"crossref","unstructured":"Farzaneh, A.H., Qi, X.: Discriminant distribution-agnostic loss for facial expression recognition in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 406\u2013407 (2020)","DOI":"10.1109\/CVPRW50498.2020.00211"},{"key":"1628_CR34","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., et al.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans on Image Process 29, 4057\u20134069 (2020)","journal-title":"IEEE Trans on Image Process"},{"key":"1628_CR35","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"1628_CR36","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., et al.: Cbam: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"1628_CR37","doi-asserted-by":"crossref","unstructured":"Bello, I., Zoph, B., Vaswani, A., et al.: Attention augmented convolutional networks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3286\u20133295 (2019)","DOI":"10.1109\/ICCV.2019.00338"},{"key":"1628_CR38","doi-asserted-by":"crossref","unstructured":"Chen, Y., Rohrbach, M., Yan, Z., et al.: Graph-based global reasoning networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 433\u2013442 (2019)","DOI":"10.1109\/CVPR.2019.00052"},{"key":"1628_CR39","doi-asserted-by":"publisher","first-page":"428","DOI":"10.1016\/j.neucom.2020.03.111","volume":"407","author":"J Cai","year":"2020","unstructured":"Cai, J., Hu, J., Tang, X., et al.: Deep historical long short-term memory network for action recognition. Neurocomputing 407, 428\u2013438 (2020). https:\/\/doi.org\/10.1016\/j.neucom.2020.03.111","journal-title":"Neurocomputing"},{"key":"1628_CR40","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2024.3384333","author":"J Cai","year":"2024","unstructured":"Cai, J., Li, Y., Liu, B., et al.: Developing deep LSTMs with later temporal attention for predicting COVID-19 severity, clinical outcome, and antibody level by screening serological indicators over time. IEEE J Biomed Health (2024). https:\/\/doi.org\/10.1109\/JBHI.2024.3384333","journal-title":"IEEE J Biomed Health"},{"key":"1628_CR41","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1016\/j.neucom.2018.03.034","volume":"296","author":"W Sun","year":"2018","unstructured":"Sun, W., Zhao, H., Jin, Z.: A visual attention based ROI detection method for facial expression recognition. Neurocomputing 296, 12\u201322 (2018). https:\/\/doi.org\/10.1016\/j.neucom.2018.03.034","journal-title":"Neurocomputing"},{"key":"1628_CR42","doi-asserted-by":"crossref","unstructured":"Marrero Fernandez, P.D., Guerrero Pena, F.A., Ren, T., et al.: Feratt: facial expression recognition with attention net. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 1\u201310 (2019)","DOI":"10.1109\/CVPRW.2019.00112"},{"key":"1628_CR43","doi-asserted-by":"crossref","unstructured":"Gao, Z., Xie, J., Wang, Q., et al.: Global second-order pooling convolutional networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3024\u20133033 (2019)","DOI":"10.1109\/CVPR.2019.00314"},{"key":"1628_CR44","doi-asserted-by":"crossref","unstructured":"Farzaneh, A.H., Qi, X.: Facial expression recognition in the wild via deep attentive center loss. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2402\u20132411 (2021)","DOI":"10.1109\/WACV48630.2021.00245"},{"key":"1628_CR45","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13713\u201313722 (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"1628_CR46","doi-asserted-by":"crossref","unstructured":"Feng, L., Shu, S., Lin, Z., et al.: Can the cross-entropy loss be robust to label noise? In: Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pp. 2206\u20132212 (2021)","DOI":"10.24963\/ijcai.2020\/305"},{"key":"1628_CR47","doi-asserted-by":"crossref","unstructured":"Gonzalez, S., Miikkulainen, R.: Optimizing loss functions through multi-variate taylor polynomial parameterization. In: Proceedings of the Genetic and Evolutionary Computation Conference, pp. 305\u2013313 (2021)","DOI":"10.1145\/3449639.3459277"},{"key":"1628_CR48","unstructured":"Leng, Z., Tan, M., Liu, C., et al.: PolyLoss: a polynomial expansion perspective of classification loss functions. In: Proceedings of the International Conference on Learning Representations, pp. 1\u201316 (2021)"},{"key":"1628_CR49","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/TIP.2018.2868382","volume":"28","author":"S Li","year":"2018","unstructured":"Li, S., Deng, W.: Reliable crowdsourcing and deep locality-preserving learning for unconstrained facial expression recognition. IEEE Trans. Image Process. 28, 356\u2013370 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"1628_CR50","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: Affectnet: A database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18\u201331 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1628_CR51","unstructured":"Li, Y., Lu, Y., Li, J., et al.: Separate loss for basic and compound facial expression recognition in the wild. In: Proceedings of the Asian Conference on Machine Learning (PMLR), pp. 897\u2013911 (2019)"},{"issue":"5","key":"1628_CR52","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/TIP.2018.2886767","volume":"28","author":"Y Li","year":"2018","unstructured":"Li, Y., Zeng, J., Shan, S., et al.: Occlusion aware facial expression recognition using CNN with attention mechanism. IEEE Trans. Image Process. 28(5), 2439\u20132450 (2018). https:\/\/doi.org\/10.1109\/TIP.2018.2886767","journal-title":"IEEE Trans. Image Process."},{"key":"1628_CR53","unstructured":"Arora, S., Hu, W. and Kothari, P.K.: An analysis of the t-sne algorithm for data visualization. In: Proceedings of the Asian Conference on Machine Learning (PMLR), pp. 1455\u20131462 (2018)"},{"issue":"10","key":"1628_CR54","doi-asserted-by":"publisher","first-page":"2905","DOI":"10.1007\/s10994-020-05917-0","volume":"110","author":"B Kang","year":"2021","unstructured":"Kang, B., Garc\u00eda Garc\u00eda, D., Lijffijt, J., et al.: Conditional t-SNE: more informative t-SNE embeddings. Mach. Learn. 110(10), 2905\u20132940 (2021)","journal-title":"Mach. Learn."},{"issue":"1","key":"1628_CR55","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1109\/TPAMI.2022.3152247","volume":"45","author":"K Han","year":"2022","unstructured":"Han, K., Wang, Y., Chen, H., et al.: A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 87\u2013110 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1628_CR56","doi-asserted-by":"crossref","unstructured":"Xue, F., Wang, Q. and Guo, G.: Transfer: learning relation-aware facial expression representations with transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3601\u20133610 (2021)","DOI":"10.1109\/ICCV48922.2021.00358"},{"key":"1628_CR57","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2021.3122146","author":"F Ma","year":"2021","unstructured":"Ma, F., Sun, B., Li, S.: Facial expression recognition with visual transformers and attentional selective fusion. IEEE Trans. Affect. Comput. (Early Access) (2021). https:\/\/doi.org\/10.1109\/TAFFC.2021.3122146","journal-title":"IEEE Trans. Affect. Comput. (Early Access)"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01628-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01628-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01628-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,28]],"date-time":"2025-02-28T11:05:00Z","timestamp":1740740700000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01628-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,3]]},"references-count":57,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["1628"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01628-x","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,3]]},"assertion":[{"value":"27 May 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All authors read and approved the final version of the manuscript.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"All authors contributed to this work.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"All authors have checked the manuscript and have agreed to the submission.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}],"article-number":"43"}}