{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T17:00:06Z","timestamp":1775667606366,"version":"3.50.1"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2022,3,4]],"date-time":"2022-03-04T00:00:00Z","timestamp":1646352000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,3,4]],"date-time":"2022-03-04T00:00:00Z","timestamp":1646352000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61966035"],"award-info":[{"award-number":["61966035"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the Autonomous Region Science and Technology De-partment International Cooperation Project","award":["2020E01023"],"award-info":[{"award-number":["2020E01023"]}]},{"name":"Tianshan Innovation Team Plan Project of Xinjiang Uy-gur Autonomous Region","award":["202101642"],"award-info":[{"award-number":["202101642"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2022,6]]},"DOI":"10.1007\/s00521-022-07016-8","type":"journal-article","created":{"date-parts":[[2022,3,4]],"date-time":"2022-03-04T19:02:26Z","timestamp":1646420546000},"page":"10175-10187","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Effective attention feature reconstruction loss for facial expression recognition in the wild"],"prefix":"10.1007","volume":"34","author":[{"given":"Weijun","family":"Gong","sequence":"first","affiliation":[]},{"given":"Yingying","family":"Fan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6564-4745","authenticated-orcid":false,"given":"Yurong","family":"Qian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,3,4]]},"reference":[{"key":"7016_CR1","doi-asserted-by":"crossref","unstructured":"Acharya D, Huang Z, Pani Paudel D, Van Gool L (2018) Covariance pooling for facial expression recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW),\u00a0pp 367\u2013374. IEEE","DOI":"10.1109\/CVPRW.2018.00077"},{"key":"7016_CR2","doi-asserted-by":"crossref","unstructured":"Albanie S, Nagrani A, Vedaldi A, Zisserman A (2018) Emotion recognition in speech using cross-modal transfer in the wild. In: Proceedings of the 26th ACM international conference on Multimedia, pp 292\u2013301","DOI":"10.1145\/3240508.3240578"},{"key":"7016_CR3","doi-asserted-by":"crossref","unstructured":"Barsoum E, Zhang C, Ferrer CC, Zhang Z (2016) Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction (ICMI), pp 279\u2013283","DOI":"10.1145\/2993148.2993165"},{"key":"7016_CR4","doi-asserted-by":"crossref","unstructured":"Cai J, Meng Z, Khan AS, Li Z, O'Reilly J, Tong Y (2018) Island loss for learning discriminative features in facial expression recognition. In: 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp 302\u2013309. IEEE","DOI":"10.1109\/FG.2018.00051"},{"key":"7016_CR5","doi-asserted-by":"crossref","unstructured":"Fan X, Deng Z, Wang K, Peng X, Qiao Y (2020) Learning discriminative representation for facial expression recognition from uncertainties. In: 2020 IEEE International Conference on Image Processing (ICIP), pp 903\u2013907. IEEE","DOI":"10.1109\/ICIP40778.2020.9190643"},{"key":"7016_CR6","doi-asserted-by":"crossref","unstructured":"Farzaneh AH, Qi X (2020) Discriminant Distribution-Agnostic Loss for Facial Expression Recognition in the Wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp 406\u2013407. IEEE","DOI":"10.1109\/CVPRW50498.2020.00211"},{"key":"7016_CR7","doi-asserted-by":"publisher","first-page":"64827","DOI":"10.1109\/ACCESS.2019.2917266","volume":"7","author":"MI Georgescu","year":"2019","unstructured":"Georgescu MI, Ionescu RT, Popescu M (2019) Local learning with deep and handcrafted features for facial expression recognition. IEEE Access 7:64827\u201364836","journal-title":"IEEE Access"},{"key":"7016_CR8","doi-asserted-by":"crossref","unstructured":"Goodfellow IJ, Erhan D, Carrier PL, Courville A, Mirza M, Hamner B, Bengio Y (2013) Challenges in representation learning: A report on three machine learning contests. In: International Conference on Neural Information Processing, pp 117\u2013124. Springer","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"7016_CR9","doi-asserted-by":"crossref","unstructured":"Guo Y, Zhang L, Hu Y, He X, Gao J (2016) Ms-celeb-1m: A dataset and benchmark for large-scale face recognition. In: Proceedings of the European conference on computer vision (ECCV), pp 87\u2013102. Springer","DOI":"10.1007\/978-3-319-46487-9_6"},{"key":"7016_CR10","doi-asserted-by":"crossref","unstructured":"Hou Q, Zhou D, Feng J (2021) Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 13713\u201313722. IEEE","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"7016_CR11","doi-asserted-by":"crossref","unstructured":"Huang C (2017) Combining convolutional neural networks for emotion recognition. In: 2017 IEEE MIT Undergraduate Research Technology Conference (URTC), pp 1\u20134. IEEE","DOI":"10.1109\/URTC.2017.8284175"},{"key":"7016_CR12","first-page":"1","volume":"70","author":"M Karnati","year":"2020","unstructured":"Karnati M, Seal A, Krejcar O, Yazidi A (2020) Facial expression recognition using local gravitational force descriptor-based deep convolution neural networks. IEEE Trans Instrum Meas 70:1\u201312","journal-title":"IEEE Trans Instrum Meas"},{"key":"7016_CR13","doi-asserted-by":"publisher","first-page":"9125","DOI":"10.1007\/s00521-020-05676-y","volume":"33","author":"M Karnati","year":"2021","unstructured":"Karnati M, Seal A, Krejcar O, Yazidi A (2021) FER-net: facial expression recognition using deep neural net. Neural Comput Appl 33:9125\u20139136","journal-title":"Neural Comput Appl"},{"key":"7016_CR14","doi-asserted-by":"crossref","unstructured":"Karnati M, Seal A, Yazidi A, Krejcar O (2021) LieNet: a deep convolution neural networks framework for detecting deception. IEEE Trans Cogn Develop Syst. 126(5): 550\u2013569","DOI":"10.1109\/TCDS.2021.3086011"},{"key":"7016_CR15","doi-asserted-by":"crossref","unstructured":"Li Y, Zeng J, Shan S, Chen X (2018) Patch-gated CNN for occlusion-aware facial expression recognition. In: 2018 24th International Conference on Pattern Recognition (ICPR), pp 2209\u20132214. IEEE","DOI":"10.1109\/ICPR.2018.8545853"},{"key":"7016_CR16","unstructured":"Li Y, Lu Y, Li J, Lu G (2019) Separate loss for basic and compound facial expression recognition in the wild. In: Asian Conference on Machine Learning (ACML), pp 897\u2013911"},{"key":"7016_CR17","doi-asserted-by":"crossref","unstructured":"Li S, Deng W, Du JP (2017) Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE Conference On Computer Vision And Pattern Recognition (CVPR), pp 2852\u20132861. IEEE","DOI":"10.1109\/CVPR.2017.277"},{"issue":"5","key":"7016_CR18","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/TIP.2018.2886767","volume":"28","author":"Y Li","year":"2018","unstructured":"Li Y, Zeng J, Shan S, Chen X (2018) Occlusion aware facial expression recognition using CNN with attention mechanism. IEEE Trans Image Process 28(5):2439\u20132450","journal-title":"IEEE Trans Image Process"},{"key":"7016_CR19","doi-asserted-by":"publisher","first-page":"2016","DOI":"10.1109\/TIP.2021.3049955","volume":"30","author":"H Li","year":"2021","unstructured":"Li H, Wang N, Ding X, Yang X, Gao X (2021) Adaptively learning facial expression representation via CF labels and distillation. IEEE Trans Image Process 30:2016\u20132028","journal-title":"IEEE Trans Image Process"},{"key":"7016_CR20","doi-asserted-by":"crossref","unstructured":"Lin TY, Goyal P, Girshick R, He K, Doll\u00e1r P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp 2980\u20132988. IEEE","DOI":"10.1109\/ICCV.2017.324"},{"key":"7016_CR21","doi-asserted-by":"crossref","unstructured":"Liu W, Wen Y, Yu Z, Li M, Raj B, Song L (2017) Sphereface: Deep hypersphere embedding for face recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 212\u2013220. IEEE","DOI":"10.1109\/CVPR.2017.713"},{"key":"7016_CR22","doi-asserted-by":"crossref","unstructured":"Lucey P, Cohn JF, Kanade T, Saragih J, Ambadar Z, Mat thews I (2010) The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference On Computer Vision And Pattern Recognition-Workshops (CVPRW), pp 94\u2013101. IEEE","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"7016_CR23","unstructured":"Lyons M, Akamatsu S, Kamachi M, Gyoba J (1998) Coding facial expressions with gabor wavelets. In: Third IEEE International Conference on Automatic Face and Gesture Recognition (FG), pp 200\u2013205. IEEE"},{"key":"7016_CR24","unstructured":"Misra D (2019) Mish: A self regularized non-monotonic neural activation function. arXiv preprint arXiv:1908.08681"},{"issue":"1","key":"7016_CR25","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini A, Hasani B, Mahoor MH (2017) Affectnet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans Affect Comput 10(1):18\u201331","journal-title":"IEEE Trans Affect Comput"},{"key":"7016_CR26","unstructured":"M\u00fcller R, Kornblith S, Hinton G (2019) When does label smoothing help? In: Proceedings of the 33rd International Conference on Neural Information Processing Systems (NIPS), pp 4694\u20134703"},{"key":"7016_CR27","doi-asserted-by":"crossref","unstructured":"Ou J, Bai XB, Pei Y, Ma L, Liu W (2010) Automatic facial expression recognition using Gabor filter and expression analysis. In: 2010 Second International Conference on Computer Modeling and Simulation (ICCMS), pp 215\u2013218. IEEE","DOI":"10.1109\/ICCMS.2010.45"},{"key":"7016_CR28","doi-asserted-by":"crossref","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2017) Grad-cam: Visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE international conference on computer vision (ICCV), pp 618\u2013626. IEEE","DOI":"10.1109\/ICCV.2017.74"},{"issue":"6","key":"7016_CR29","doi-asserted-by":"publisher","first-page":"803","DOI":"10.1016\/j.imavis.2008.08.005","volume":"27","author":"C Shan","year":"2009","unstructured":"Shan C, Gong S, McOwan PW (2009) Facial expression recognition based on local binary patterns: a comprehensive study. Image Vis Comput 27(6):803\u2013816","journal-title":"Image Vis Comput"},{"key":"7016_CR30","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.neucom.2019.05.005","volume":"355","author":"J Shao","year":"2019","unstructured":"Shao J, Qian Y (2019) Three convolutional neural network models for facial expression recognition in the wild. Neurocomputing 355:82\u201392","journal-title":"Neurocomputing"},{"key":"7016_CR31","doi-asserted-by":"crossref","unstructured":"Siqueira H, Magg S, Wermter S (2020) Efficient facial feature learning with wide ensemble-based convolutional neural networks. In: Proceedings of the AAAI conference on artificial intelligence (AAAI), pp 5800\u20135809","DOI":"10.1609\/aaai.v34i04.6037"},{"issue":"11","key":"7016_CR32","first-page":"2579","volume":"9","author":"L Van der Maaten","year":"2008","unstructured":"Van der Maaten L, Hinton G (2008) Visualizing data using t-SNE. J Mach Learn Res 9(11):2579\u20132605","journal-title":"J Mach Learn Res"},{"key":"7016_CR33","doi-asserted-by":"publisher","first-page":"131988","DOI":"10.1109\/ACCESS.2020.3010018","volume":"8","author":"TH Vo","year":"2020","unstructured":"Vo TH, Lee GS, Yang HJ, Kim SH (2020) Pyramid with super resolution for in-the-wild facial expression recognition. IEEE Access 8:131988\u2013132001","journal-title":"IEEE Access"},{"key":"7016_CR34","doi-asserted-by":"crossref","unstructured":"Wang K, Peng X, Yang J, Lu S, Qiao Y (2020) Suppressing uncertainties for large-scale facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 6897\u20136906. IEEE","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"7016_CR35","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang K, Peng X, Yang J, Meng D, Qiao Y (2020) Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans Image Process 29:4057\u20134069","journal-title":"IEEE Trans Image Process"},{"key":"7016_CR36","doi-asserted-by":"crossref","unstructured":"Wen Y, Zhang K, Li Z, Qiao Y (2016) A discriminative feature learning approach for deep face recognition. In: European conference on computer vision (ECCV), pp 499\u2013515. Springer","DOI":"10.1007\/978-3-319-46478-7_31"},{"issue":"2","key":"7016_CR37","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1109\/MMUL.2021.3076834","volume":"28","author":"HY Xia","year":"2021","unstructured":"Xia HY, Li C, Tan Y, Li L, Song S (2021) Destruction and reconstruction learning for facial expression recognition. IEEE Multimedia 28(2):20\u201328","journal-title":"IEEE Multimedia"},{"key":"7016_CR38","doi-asserted-by":"crossref","unstructured":"Yu Z, Zhang C (2015) Image based static facial expression recognition with multiple deep network learning. In: Proceedings of the 2015 ACM on international conference on multimodal interaction (ICMI), pp 435\u2013442","DOI":"10.1145\/2818346.2830595"},{"key":"7016_CR39","doi-asserted-by":"crossref","unstructured":"Zeng J, Shan S, Chen X (2018) Facial expression recognition with inconsistently annotated datasets. In: Proceedings of the European conference on computer vision (ECCV), pp 222\u2013237","DOI":"10.1007\/978-3-030-01261-8_14"},{"issue":"5","key":"7016_CR40","doi-asserted-by":"publisher","first-page":"550","DOI":"10.1007\/s11263-017-1055-1","volume":"126","author":"Z Zhang","year":"2018","unstructured":"Zhang Z, Luo P, Loy CC, Tang X (2018) From facial expression recognition to interpersonal relation prediction. Int J Comput Vis 126(5):550\u2013569","journal-title":"Int J Comput Vis"},{"issue":"6","key":"7016_CR41","doi-asserted-by":"publisher","first-page":"915","DOI":"10.1109\/TPAMI.2007.1110","volume":"29","author":"G Zhao","year":"2007","unstructured":"Zhao G, Pietikainen M (2007) Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans Pattern Anal Mach Intell 29(6):915\u2013928","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"7016_CR42","doi-asserted-by":"crossref","unstructured":"Zhao Z, Liu Q, Zhou F (2021) Robust lightweight facial expression recognition network with label distribution training. In: Proceedings of the AAAI Conference on Artificial Intelligence. 35(4), pp 3510\u20133519","DOI":"10.1609\/aaai.v35i4.16465"},{"issue":"9","key":"7016_CR43","doi-asserted-by":"publisher","first-page":"607","DOI":"10.1016\/j.imavis.2011.07.002","volume":"29","author":"G Zhao","year":"2011","unstructured":"Zhao G, Huang X, Taini M, Li SZ, Pietik\u00e4Inen M (2011) Facial expression recognition from near-infrared videos. Image Vis Comput 29(9):607\u2013619","journal-title":"Image Vis Comput"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-07016-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-022-07016-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-07016-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,28]],"date-time":"2023-01-28T11:37:01Z","timestamp":1674905821000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-022-07016-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,3,4]]},"references-count":43,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2022,6]]}},"alternative-id":["7016"],"URL":"https:\/\/doi.org\/10.1007\/s00521-022-07016-8","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,3,4]]},"assertion":[{"value":"7 October 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 January 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 March 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We wish to submit a new manuscript entitled \u201cEffective Attention Feature Reconstruction Loss for\u00a0Facial Expression Recognition in the Wild\u201d for consideration in the Neural Computing and Applications Journal. We declare that this work is original and there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}