{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,18]],"date-time":"2026-02-18T23:46:59Z","timestamp":1771458419861,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2022,8,10]],"date-time":"2022-08-10T00:00:00Z","timestamp":1660089600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,8,10]],"date-time":"2022-08-10T00:00:00Z","timestamp":1660089600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62006049"],"award-info":[{"award-number":["62006049"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62172113"],"award-info":[{"award-number":["62172113"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176095"],"award-info":[{"award-number":["62176095"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62072123"],"award-info":[{"award-number":["62072123"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Guangdong Province Key Area R&D Plan Project","award":["2020B1111120001"],"award-info":[{"award-number":["2020B1111120001"]}]},{"DOI":"10.13039\/501100010256","name":"Guangzhou Municipal Science and Technology Project","doi-asserted-by":"publisher","award":["201803010088"],"award-info":[{"award-number":["201803010088"]}],"id":[{"id":"10.13039\/501100010256","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ministry of Education Humanities and Social Science project","award":["18JDGC012"],"award-info":[{"award-number":["18JDGC012"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s00371-022-02619-7","type":"journal-article","created":{"date-parts":[[2022,8,10]],"date-time":"2022-08-10T19:25:20Z","timestamp":1660159520000},"page":"4709-4720","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Heuristic objective for facial expression recognition"],"prefix":"10.1007","volume":"39","author":[{"given":"Huihui","family":"Li","sequence":"first","affiliation":[]},{"given":"Xiangling","family":"Xiao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0795-841X","authenticated-orcid":false,"given":"Xiaoyong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jianhua","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Guihua","family":"Wen","sequence":"additional","affiliation":[]},{"given":"Peng","family":"Liang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,10]]},"reference":[{"key":"2619_CR1","doi-asserted-by":"publisher","first-page":"397","DOI":"10.1016\/j.neucom.2016.12.043","volume":"230","author":"Y Sun","year":"2017","unstructured":"Sun, Y., Wen, G.: Cognitive facial expression recognition with constrained dimensionality reduction. Neurocomputing 230, 397\u2013408 (2017)","journal-title":"Neurocomputing"},{"issue":"1","key":"2619_CR2","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2019","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: AffectNet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18\u201331 (2019)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"2619_CR3","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE conference on computer vision pattern recognition (CVPR), pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"2619_CR4","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Lucey, S., Gedeon, T.: Static facial expression analysis in tough conditions: data, evaluation protocol and benchmark. In: 2011 IEEE international conference on computer vision workshops (ICCVW) (2011)","DOI":"10.1109\/ICCVW.2011.6130508"},{"key":"2619_CR5","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1007\/978-3-642-42051-1_16","volume-title":"Neural information processing","author":"IJ Goodfellow","year":"2013","unstructured":"Goodfellow, I.J., Erhan, D., Carrier, P.L., Courville, A., Mirza, M., Hamner, B., Cukierski, W., Tang, Y., Thaler, D., Lee, D.H., et al.: Challenges in representation learning: a report on three machine learning contests. In: Lee, M., Hirose, A., Hou, Z.G., Kil, R.M. (eds.) Neural information processing, pp. 117\u2013124. Springer, Berlin (2013). https:\/\/doi.org\/10.1007\/978-3-642-42051-1_16"},{"key":"2619_CR6","doi-asserted-by":"crossref","unstructured":"Barsoum, E., Zhang, C., Ferrer C.C., Zhang, Z.: Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM international conference on multimodal interaction, pp. 279\u2013283 (2016)","DOI":"10.1145\/2993148.2993165"},{"key":"2619_CR7","doi-asserted-by":"crossref","unstructured":"Fabian Benitez-Quiroz, C., Srinivasan, R., Martinez, A.M.: Emotionet: an accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.600"},{"key":"2619_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2017\/1945630","volume":"2017","author":"G Wen","year":"2017","unstructured":"Wen, G., Li, H., Huang, J., et al.: Random deep belief networks for recognizing emotions from speech signals. Comput. Intell. Neurosci. 2017, 1\u20139 (2017). https:\/\/doi.org\/10.1155\/2017\/1945630","journal-title":"Comput. Intell. Neurosci."},{"key":"2619_CR9","doi-asserted-by":"crossref","unstructured":"Chen, S., Wang, J., Chen, Y., Shi, Z., Geng, X., Rui, Y.: Label distribution learning on auxiliary label space graphs for facial expression recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (2020)","DOI":"10.1109\/CVPR42600.2020.01400"},{"issue":"07","key":"2619_CR10","doi-asserted-by":"publisher","first-page":"1311","DOI":"10.1109\/TLA.2020.9099774","volume":"18","author":"MG Villanueva","year":"2020","unstructured":"Villanueva, M.G., Zavala, S.R.: Deep neural network architecture: application for facial expression recognition. IEEE Latin Am. Trans. 18(07), 1311\u20131319 (2020). https:\/\/doi.org\/10.1109\/TLA.2020.9099774","journal-title":"IEEE Latin Am. Trans."},{"key":"2619_CR11","doi-asserted-by":"publisher","unstructured":"Joseph, J.L., Mathew, S.P.: Facial expression recognition for the blind using deep learning. In: 2021 IEEE 4th international conference on computing, power and communication technologies (GUCON), pp. 1\u20135 (2021). https:\/\/doi.org\/10.1109\/GUCON50781.2021.9574035.","DOI":"10.1109\/GUCON50781.2021.9574035"},{"key":"2619_CR12","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-022-02483-5","author":"C Liu","year":"2022","unstructured":"Liu, C., Liu, X., Chen, C., Wang, Q.: Soft thresholding squeeze-and-excitation network for pose-invariant facial expression recognition. Visual Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-022-02483-5","journal-title":"Visual Comput."},{"key":"2619_CR13","unstructured":"Jun Wang. Facial Expression and Action Unit Recognition Based on Prior Knowledge. University of Science and Technology of China, 2015."},{"key":"2619_CR14","volume-title":"What the Face Reveals: Basic and Applied Studies of Spontaneous Expression using The Facial Action Coding System (FACS)","author":"P Ekman","year":"1997","unstructured":"Ekman, P., Rosenberg, E.L.: What the Face Reveals: Basic and Applied Studies of Spontaneous Expression using The Facial Action Coding System (FACS). Oxford University Press, USA (1997)"},{"key":"2619_CR15","doi-asserted-by":"publisher","unstructured":"Nuanes, T., Elsey, M., Sankaranarayanan, A., Shen, J.: Soft cross entropy loss and bottleneck tri-cost volume for efficient stereo depth prediction. In: 2021 IEEE\/CVF conference on computer vision and pattern recognition workshops (CVPRW), pp. 2840-2848 (2021). doi: https:\/\/doi.org\/10.1109\/CVPRW53098.2021.00319","DOI":"10.1109\/CVPRW53098.2021.00319"},{"key":"2619_CR16","volume-title":"Machine Learning: An Artificial Intelligence Approach","author":"JR Anderson","year":"1986","unstructured":"Anderson, J.R., Michalski, R.S., Carbonell, J.G., et al.: Machine Learning: An Artificial Intelligence Approach. Morgan Kaufmann, Burlington (1986)"},{"issue":"4","key":"2619_CR17","doi-asserted-by":"publisher","first-page":"898","DOI":"10.1109\/TCDS.2020.3034807","volume":"13","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Su, W., Yu, J., Wang, Z.: Identity\u2013expression dual branch network for facial expression recognition. IEEE Trans. Cognitive Dev. Syst. 13(4), 898\u2013911 (2021). https:\/\/doi.org\/10.1109\/TCDS.2020.3034807","journal-title":"IEEE Trans. Cognitive Dev. Syst."},{"issue":"2","key":"2619_CR18","doi-asserted-by":"publisher","first-page":"544","DOI":"10.1109\/TAFFC.2018.2880201","volume":"12","author":"M Li","year":"2021","unstructured":"Li, M., Hao, X., Huang, X., Song, Z., Liu, X., Li, X.: Facial expression recognition with identity and emotion joint learning. IEEE Trans. Affective Comput. 12(2), 544\u2013550 (2021). https:\/\/doi.org\/10.1109\/TAFFC.2018.2880201","journal-title":"IEEE Trans. Affective Comput."},{"issue":"1","key":"2619_CR19","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1109\/TII.2021.3075989","volume":"18","author":"J Chen","year":"2022","unstructured":"Chen, J., Guo, C., Xu, R., Zhang, K., Yang, Z., Liu, H.: Toward children\u2019s empathy ability analysis: joint facial expression recognition and intensity estimation using label distribution learning. IEEE Trans. Industr. Inf. 18(1), 16\u201325 (2022). https:\/\/doi.org\/10.1109\/TII.2021.3075989","journal-title":"IEEE Trans. Industr. Inf."},{"issue":"2","key":"2619_CR20","doi-asserted-by":"publisher","first-page":"544","DOI":"10.1109\/TKDE.2020.2985365","volume":"34","author":"T Zhang","year":"2022","unstructured":"Zhang, T., et al.: Cross-database micro-expression recognition: a benchmark. IEEE Trans. Knowl. Data Eng. 34(2), 544\u2013559 (2022). https:\/\/doi.org\/10.1109\/TKDE.2020.2985365","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"2619_CR21","unstructured":"Wang, J., et al.: Capture expression-dependent AU relations for expression recognition. In: 2014 IEEE international conference on multimedia and expo workshops (ICMEW) IEEE (2014)"},{"key":"2619_CR22","doi-asserted-by":"crossref","unstructured":"Wang, Z., Chen, T., Ren, J., Yu, W., Cheng, H., Lin, L.: Deep reasoning with knowledge graph for social relationship understanding. In Proceedings of the international joint conference on artificial intelligence, pp. 2021\u20132028 (2018)","DOI":"10.24963\/ijcai.2018\/142"},{"key":"2619_CR23","doi-asserted-by":"crossref","unstructured":"Chen, T., Yu, W., Chen, R., Lin, L.: Knowledge-embedded routing network for scene graph generation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6163\u20136171 (2019)","DOI":"10.1109\/CVPR.2019.00632"},{"key":"2619_CR24","doi-asserted-by":"crossref","unstructured":"Chen, T., Xu, M., Hui, X., Wu, H., Lin, L.: Learning semanticspecific graph representation for multi-label image recognition. In: Proceedings of the IEEE international conference on computer vision, pp. 522\u2013531 (2019)","DOI":"10.1109\/ICCV.2019.00061"},{"key":"2619_CR25","doi-asserted-by":"crossref","unstructured":"Xie, Y., Chen, T., Pu, T., Wu, H., Lin, L.: Adversarial graph representation adaptation for cross-domain facial expression recognition. In: Proceedings of the 28th ACM international conference on multimedia, pp. 1255\u20131264 (2020)","DOI":"10.1145\/3394171.3413822"},{"key":"2619_CR26","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3131222","author":"T Chen","year":"2020","unstructured":"Chen, T., Pu, T., Xie, Y., Wu, H., Liu, L., Lin, L.: Cross-domain facial expression recognition: a unified evaluation benchmark and adversarial graph learning. IEEE Trans Pattern Anal. Mach. Intell. (2020). https:\/\/doi.org\/10.1109\/TPAMI.2021.3131222","journal-title":"IEEE Trans Pattern Anal. Mach. Intell."},{"key":"2619_CR27","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3025814","author":"T Chen","year":"2020","unstructured":"Chen, T., Lin, L., Hui, X., Chen, R., Wu, H.: Knowledge-guided multi-label few-shot learning for general image recognition. IEEE Trans. Pattern Anal. Mach. Intell. (2020). https:\/\/doi.org\/10.1109\/TPAMI.2020.3025814","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2619_CR28","doi-asserted-by":"publisher","unstructured":"Pu, T., Chen, T., Xie, Y., Wu H., Lin, L.: AU-expression knowledge constrained representation learning for facial expression recognition. In: 2021 IEEE international conference on robotics and automation (ICRA), pp. 11154-11161 (2021). https:\/\/doi.org\/10.1109\/ICRA48506.2021.9561252.","DOI":"10.1109\/ICRA48506.2021.9561252"},{"issue":"4","key":"2619_CR29","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1007\/s12193-020-00363-7","volume":"15","author":"J He","year":"2021","unstructured":"He, J., Xiaocui, Y., Sun, B., Lejun, Y.: Facial expression and action unit recognition augmented by their dependencies on graph convolutional networks. J. Multimodal User Interfaces 15(4), 429\u2013440 (2021). https:\/\/doi.org\/10.1007\/s12193-020-00363-7","journal-title":"J. Multimodal User Interfaces"},{"key":"2619_CR30","doi-asserted-by":"publisher","first-page":"7143","DOI":"10.1109\/TIP.2021.3101820","volume":"30","author":"X Jin","year":"2021","unstructured":"Jin, X., Lai, Z., Jin, Z.: Learning dynamic relationships for facial expression recognition based on graph convolutional network. IEEE Trans. Image Process. 30, 7143\u20137155 (2021). https:\/\/doi.org\/10.1109\/TIP.2021.3101820","journal-title":"IEEE Trans. Image Process."},{"issue":"11","key":"2619_CR31","doi-asserted-by":"publisher","first-page":"2914","DOI":"10.1109\/TMM.2020.2966858","volume":"22","author":"G Wen","year":"2020","unstructured":"Wen, G., Chang, T., Li, H., Jiang, L.: Dynamic objectives learning for facial expression recognition. IEEE Trans. Multimedia 22(11), 2914\u20132925 (2020). https:\/\/doi.org\/10.1109\/TMM.2020.2966858","journal-title":"IEEE Trans. Multimedia"},{"key":"2619_CR32","doi-asserted-by":"crossref","unstructured":"Pan, H., Han, H., Shan, S., Chen, X.: Mean-variance loss for deep age estimation from a face. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 5285\u20135294 (2018)","DOI":"10.1109\/CVPR.2018.00554"},{"key":"2619_CR33","doi-asserted-by":"publisher","first-page":"499","DOI":"10.1007\/978-3-319-46478-7_31","volume-title":"Computer vision \u2013 ECCV 2016: 14th european conference, Amsterdam, The Netherlands, October 11\u201314, 2016, proceedings, Part VII","author":"Y Wen","year":"2016","unstructured":"Wen, Y., Zhang, K., Li, Y., Qiao, Y.: A discriminative feature learning approach for deep face recognition. In: Leibe, Bastian, Matas, Jiri, Sebe, Nicu, Welling, Max (eds.) Computer vision \u2013 ECCV 2016: 14th european conference, Amsterdam, The Netherlands, October 11\u201314, 2016, proceedings, Part VII, pp. 499\u2013515. Springer International Publishing, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46478-7_31"},{"key":"2619_CR34","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-022-02429-x","author":"M Sandhya","year":"2022","unstructured":"Sandhya, M., Morampudi, M.K., Pruthweraaj, I., et al.: Multi-instance cancelable iris authentication system using triplet loss for deep learning models. Vis. Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-022-02429-x","journal-title":"Vis. Comput."},{"key":"2619_CR35","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"7","key":"2619_CR36","doi-asserted-by":"publisher","first-page":"926","DOI":"10.1109\/LSP.2018.2822810","volume":"25","author":"F Wang","year":"2018","unstructured":"Wang, F., Cheng, J., Liu, W., Liu, H.: Additive margin softmax for face verification. IEEE Signal Process. Lett. 25(7), 926\u2013930 (2018)","journal-title":"IEEE Signal Process. Lett."},{"key":"2619_CR37","doi-asserted-by":"crossref","unstructured":"Liu, W., Wen, Y., Yu, Z., et al.: Sphereface: deep hypersphere embedding for face recognition. In: Honolulu: IEEE conference on computer vision and pattern recognition (2017)","DOI":"10.1109\/CVPR.2017.713"},{"key":"2619_CR38","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Y., Zhou, Z., et al.: Cosface: large margin cosine loss for deep face recognition. In: Salt Lake City: IEEE conference on computer vision and pattern recognition (2018)","DOI":"10.1109\/CVPR.2018.00552"},{"key":"2619_CR39","doi-asserted-by":"crossref","unstructured":"Deng, J., Guo, J., Stefanos, Z.: Arcface: additive angular margin loss for deep face recognition. In: Seattle: IEEE\/CVF conference on computer vision and pattern recognition (2019)","DOI":"10.1109\/CVPR.2019.00482"},{"key":"2619_CR40","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2020.2981446","author":"L Shan","year":"2018","unstructured":"Shan, L., Deng, W.: Deep facial expression recognition: a survey. IEEE Trans. Affective Comput. (2018). https:\/\/doi.org\/10.1109\/TAFFC.2020.2981446","journal-title":"IEEE Trans. Affective Comput."},{"issue":"2","key":"2619_CR41","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1037\/h0030377","volume":"17","author":"P Ekman","year":"1971","unstructured":"Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion[J]. J. Pers. Soc. Psychol. 17(2), 124\u2013129 (1971)","journal-title":"J. Pers. Soc. Psychol."},{"key":"2619_CR42","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen, W.V.: Facial Action Coding System (FACS)[M]. Consulting Psychologists Press (1978)","DOI":"10.1037\/t27734-000"},{"issue":"2","key":"2619_CR43","doi-asserted-by":"publisher","first-page":"263","DOI":"10.1007\/s11760-020-01753-w","volume":"15","author":"D Zhu","year":"2020","unstructured":"Zhu, D., Tian, G., Zhu, L., Wang, W., Wang, B., Li, C.: LKRNet: a dual-branch network based on local key regions for facial expression recognition. Signal Image Video Process. 15(2), 263\u2013270 (2020). https:\/\/doi.org\/10.1007\/s11760-020-01753-w","journal-title":"Signal Image Video Process."},{"issue":"11","key":"2619_CR44","first-page":"2579","volume":"9","author":"L Van der Maaten","year":"2008","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11), 2579\u20132605 (2008)","journal-title":"J. Mach. Learn. Res."},{"key":"2619_CR45","doi-asserted-by":"crossref","unstructured":"Wang, K., Peng, X., Yang, J., Lu, S., Qiao, Y.: Suppressing uncertainties for large-scale facial expression recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 6897\u20136906 (2020)","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"2619_CR46","first-page":"17616","volume":"34","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Wang, C., Deng, W.: Relative uncertainty learning for facial expression recognition. NeurIPS 34, 17616\u201317627 (2021)","journal-title":"NeurIPS"},{"key":"2619_CR47","doi-asserted-by":"publisher","first-page":"6544","DOI":"10.1109\/TIP.2021.3093397","volume":"30","author":"Z Zhao","year":"2021","unstructured":"Zhao, Z., Liu, Q., Wang, S.: Learning deep global multi-scale and local attention features for facial expression recognition in the wild. IEEE Trans. Image Process. 30, 6544\u20136556 (2021). https:\/\/doi.org\/10.1109\/TIP.2021.3093397","journal-title":"IEEE Trans. Image Process."},{"key":"2619_CR48","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., Meng, D., Qiao, Yu.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 29, 4057\u20134069 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"2619_CR49","doi-asserted-by":"crossref","unstructured":"Ruan, D., Yan, Y., Lai, S., et al.: Feature decomposition and reconstruction learning for effective facial expression recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp. 7660\u20137669 (2021)","DOI":"10.1109\/CVPR46437.2021.00757"},{"key":"2619_CR50","doi-asserted-by":"crossref","unstructured":"She, J., Hu, Y., Shi, H., Wang, J., Shen, Q., Mei, T.: Dive into ambiguity: latent distribution mining and pairwise uncertainty estimation for facial expression recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 6248\u20136257 (2021)","DOI":"10.1109\/CVPR46437.2021.00618"},{"key":"2619_CR51","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Liu, Q., Zhou, F.: Robust lightweight facial expression recognition network with label distribution training. In: Proceedings of the AAAI conference on artificial intelligence, vol. 35, no. 4, pp. 3510\u20133519 (2021)","DOI":"10.1609\/aaai.v35i4.16465"},{"key":"2619_CR52","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2021.3122146","author":"F Ma","year":"2021","unstructured":"Ma, F., Sun, B., Li, S.: Facial expression recognition with visual transformers and attentional selective fusion. IEEE Trans. Affective Comput. (2021). https:\/\/doi.org\/10.1109\/TAFFC.2021.3122146","journal-title":"IEEE Trans. Affective Comput."},{"key":"2619_CR53","doi-asserted-by":"crossref","unstructured":"Albanie, S., Nagrani, A., Vedaldi, A., Zisserman, A.: Emotion recognition in speech using crossmodal transfer in the wild. In: Proceedings of the 26th ACM international conference on Multimedia, pp. 292\u2013301 (2018)","DOI":"10.1145\/3240508.3240578"},{"key":"2619_CR54","doi-asserted-by":"crossref","unstructured":"Lian, Z., Li, Y., Tao, J., Huang, J., Niu, M.: Region based robust facial expression analysis. In: 2018 First Asian conference on affective computing and intelligent interaction (ACII Asia), pp. 1\u20135. IEEE (2018)","DOI":"10.1109\/ACIIAsia.2018.8470391"},{"key":"2619_CR55","first-page":"71","volume":"2","author":"M Li","year":"2018","unstructured":"Li, M., Xu, H., Huang, X., Song, Z., Liu, X., Li, X.: Facial expression recognition with identity and emotion joint learning. IEEE Trans. Affect. Comput. 2, 71 (2018)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"2619_CR56","doi-asserted-by":"crossref","unstructured":"Chen, S., Wang, J., Chen, Y., Shi, Z., Geng, X., Rui, Y.: Label distribution learning on auxiliary label space graphs for facial expression recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 13984\u201313993 (2020)","DOI":"10.1109\/CVPR42600.2020.01400"},{"issue":"131","key":"2619_CR57","first-page":"988","volume":"8","author":"T-H Vo","year":"2020","unstructured":"Vo, T.-H., Lee, G.-S., Yang, H.-J., Kim, S.-H.: Pyramid with super resolution for in-the-wild facial expression recognition. IEEE Access 8(131), 988\u2013132001 (2020)","journal-title":"IEEE Access"},{"key":"2619_CR58","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1016\/j.patrec.2022.01.013","volume":"155","author":"S Darshan Gera","year":"2022","unstructured":"Darshan Gera, S., Balasubramanian, A.J.: CERN: compact facial expression recognition net. Pattern Recognit. Lett. 155, 9\u201318 (2022). https:\/\/doi.org\/10.1016\/j.patrec.2022.01.013","journal-title":"Pattern Recognit. Lett."},{"key":"2619_CR59","doi-asserted-by":"publisher","first-page":"107893","DOI":"10.1016\/j.patcog.2021.107893","volume":"115","author":"B Chen","year":"2021","unstructured":"Chen, B., Guan, W., Li, P., Ikeda, N., Hirasawa, K., Huchuan, L.: Residual multi-task learning for facial landmark localization and expression recognition. Pattern Recognit 115, 107893 (2021). https:\/\/doi.org\/10.1016\/j.patcog.2021.107893","journal-title":"Pattern Recognit"},{"key":"2619_CR60","doi-asserted-by":"crossref","unstructured":"Xue, F., Wang, Q., Guo, G.: Transfer: learning relation-aware facial expression representations with transformers. In: Proceedings of the IEEE\/CVF international conference on computer vision (ICCV), pp. 3601\u20133610 (2021)","DOI":"10.1109\/ICCV48922.2021.00358"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-022-02619-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-022-02619-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-022-02619-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,29]],"date-time":"2023-09-29T09:10:51Z","timestamp":1695978651000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-022-02619-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,10]]},"references-count":60,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["2619"],"URL":"https:\/\/doi.org\/10.1007\/s00371-022-02619-7","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,8,10]]},"assertion":[{"value":"6 July 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 August 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}