{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T04:35:19Z","timestamp":1768883719819,"version":"3.49.0"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2022,11,22]],"date-time":"2022-11-22T00:00:00Z","timestamp":1669075200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,11,22]],"date-time":"2022-11-22T00:00:00Z","timestamp":1669075200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Science Foundation of China","doi-asserted-by":"crossref","award":["61966035"],"award-info":[{"award-number":["61966035"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Science Foundation of China","doi-asserted-by":"crossref","award":["U1803261"],"award-info":[{"award-number":["U1803261"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Autonomous Region Science and Technology Department International Cooperation Project","award":["2020E01023"],"award-info":[{"award-number":["2020E01023"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2023,3]]},"DOI":"10.1007\/s00521-022-08040-4","type":"journal-article","created":{"date-parts":[[2022,11,23]],"date-time":"2022-11-23T09:11:58Z","timestamp":1669194718000},"page":"6529-6543","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["MPCSAN: multi-head parallel channel-spatial attention network for facial expression recognition in the wild"],"prefix":"10.1007","volume":"35","author":[{"given":"Weijun","family":"Gong","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6564-4745","authenticated-orcid":false,"given":"Yurong","family":"Qian","sequence":"additional","affiliation":[]},{"given":"Yingying","family":"Fan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,22]]},"reference":[{"key":"8040_CR1","unstructured":"Pantic M, Valstar M, Rademaker R, Maat L (2005) Web-based database for facial expression analysis. In: 2005 IEEE International Conference on Multimedia and Expo (ICME), pp. 5\u201315. IEEE, Amsterdam"},{"key":"8040_CR2","doi-asserted-by":"crossref","unstructured":"Lucey P, Cohn JF, Kanade T, Saragih J, Ambadar Z, Matthews I (2010) The extended cohn-kanade dataset (ck+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 94\u2013101. IEEE, San Francisco","DOI":"10.1109\/CVPRW.2010.5543262"},{"issue":"9","key":"8040_CR3","doi-asserted-by":"publisher","first-page":"607","DOI":"10.1016\/j.imavis.2011.07.002","volume":"29","author":"G Zhao","year":"2011","unstructured":"Zhao G, Huang X, Taini M, Li SZ, Pietik\u00e4Inen M (2011) Facial expression recognition from near-infrared videos. Image Vis Comput 29(9):607\u2013619","journal-title":"Image Vis Comput"},{"key":"8040_CR4","doi-asserted-by":"crossref","unstructured":"Li S, Deng W, Du J (2017) Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2852\u20132861. IEEE, Hawaii","DOI":"10.1109\/CVPR.2017.277"},{"key":"8040_CR5","doi-asserted-by":"crossref","unstructured":"Barsoum E, Zhang C, Ferrer CC, Zhang Z (2016) Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction (ICMI), pp. 279\u2013283. ACM, Tokyo","DOI":"10.1145\/2993148.2993165"},{"issue":"1","key":"8040_CR6","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini A, Hasani B, Mahoor MH (2017) Affectnet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans Affect Comput 10(1):18\u201331","journal-title":"IEEE Trans Affect Comput"},{"issue":"6","key":"8040_CR7","doi-asserted-by":"publisher","first-page":"803","DOI":"10.1016\/j.imavis.2008.08.005","volume":"27","author":"C Shan","year":"2009","unstructured":"Shan C, Gong S, McOwan PW (2009) Facial expression recognition based on local binary patterns: a comprehensive study. Image Vis Comput 27(6):803\u2013816","journal-title":"Image Vis Comput"},{"key":"8040_CR8","doi-asserted-by":"crossref","unstructured":"Hu Y, Zeng Z, Yin L, Wei X, Zhou X, Huang TS (2008) Multi-view facial expression recognition. In: 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2008), pp. 1\u20136. IEEE, Amsterdam","DOI":"10.1109\/AFGR.2008.4813445"},{"issue":"1","key":"8040_CR9","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1016\/j.patcog.2011.05.006","volume":"45","author":"W Gu","year":"2012","unstructured":"Gu W, Xiang C, Venkatesh Y, Huang D, Lin H (2012) Facial expression recognition using radial encoding of local gabor features and classifier synthesis. Pattern Recognit 45(1):80\u201391","journal-title":"Pattern Recognit"},{"key":"8040_CR10","doi-asserted-by":"crossref","unstructured":"Zhao J, Cheng Y, Xu Y, Xiong L, Li J, Zhao F, Jayashree K, Pranata S, Shen S, Xing J (2018) Towards pose invariant face recognition in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2207\u20132216. IEEE, Salt Lake City","DOI":"10.1109\/CVPR.2018.00235"},{"issue":"6","key":"8040_CR11","doi-asserted-by":"publisher","first-page":"884","DOI":"10.1007\/s11263-018-1131-1","volume":"127","author":"S Li","year":"2019","unstructured":"Li S, Deng W (2019) Blended emotion in-the-wild: multi-label facial expression recognition using crowdsourced annotations and deep locality feature learning. Int J Comput Vis 127(6):884\u2013906","journal-title":"Int J Comput Vis"},{"key":"8040_CR12","doi-asserted-by":"publisher","first-page":"107694","DOI":"10.1016\/j.patcog.2020.107694","volume":"112","author":"Z Wang","year":"2021","unstructured":"Wang Z, Zeng F, Liu S, Zeng B (2021) Oaenet: oriented attention ensemble for accurate facial expression recognition. Pattern Recognit 112:107694","journal-title":"Pattern Recognit"},{"key":"8040_CR13","doi-asserted-by":"publisher","first-page":"455","DOI":"10.1007\/s11263-021-01556-7","volume":"130","author":"D Ruan","year":"2022","unstructured":"Ruan D, Mo R, Yan Y, Chen S, Xue J, Wang H (2022) Adaptive deep disturbance-disentangled learning for facial expression recognition. Int J Comput Vis 130:455\u2013477","journal-title":"Int J Comput Vis"},{"issue":"7","key":"8040_CR14","doi-asserted-by":"publisher","first-page":"1936","DOI":"10.3390\/s20071936","volume":"20","author":"D Jeong","year":"2020","unstructured":"Jeong D, Kim BG, Dong SY (2020) Deep joint spatiotemporal network (djstn) for efficient facial expression recognition. Sensors 20(7):1936","journal-title":"Sensors"},{"key":"8040_CR15","doi-asserted-by":"crossref","unstructured":"Marrero\u00a0Fernandez PD, Guerrero\u00a0Pena FA, Ren T, Cunha A (2019) Feratt: facial expression recognition with attention net. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 0\u20130. IEEE, Long Beach","DOI":"10.1109\/CVPRW.2019.00112"},{"key":"8040_CR16","unstructured":"Fan Y, Li V, Lam JC (2020) Facial expression recognition with deeply-supervised attention network. IEEE Trans Affect Comput"},{"issue":"21","key":"8040_CR17","doi-asserted-by":"publisher","first-page":"6954","DOI":"10.3390\/s21216954","volume":"21","author":"SJ Park","year":"2021","unstructured":"Park SJ, Kim BG, Chilamkurti N (2021) A robust facial expression recognition algorithm based on multi-rate feature fusion scheme. Sensors 21(21):6954","journal-title":"Sensors"},{"key":"8040_CR18","doi-asserted-by":"crossref","unstructured":"Zeng J, Shan S, Chen X (2018) Facial expression recognition with inconsistently annotated datasets. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 222\u2013237. Springer, Munich","DOI":"10.1007\/978-3-030-01261-8_14"},{"key":"8040_CR19","unstructured":"Li Y, Lu Y, Li J, Lu G (2019) Separate loss for basic and compound facial expression recognition in the wild. In: Asian Conference on Machine Learning (ACML), pp. 897\u2013911. PMLR, Nagoya"},{"key":"8040_CR20","doi-asserted-by":"publisher","first-page":"64827","DOI":"10.1109\/ACCESS.2019.2917266","volume":"7","author":"MI Georgescu","year":"2019","unstructured":"Georgescu MI, Ionescu RT, Popescu M (2019) Local learning with deep and handcrafted features for facial expression recognition. IEEE Access 7:64827\u201364836","journal-title":"IEEE Access"},{"key":"8040_CR21","doi-asserted-by":"crossref","unstructured":"Wang K, Peng X, Yang J, Lu S, Qiao Y (2020) Suppressing uncertainties for large-scale facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6897\u20136906. IEEE, Seattle","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"8040_CR22","doi-asserted-by":"crossref","unstructured":"Farzaneh AH, Qi X (2020) Discriminant distribution-agnostic loss for facial expression recognition in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 406\u2013407. IEEE, Seattle","DOI":"10.1109\/CVPRW50498.2020.00211"},{"key":"8040_CR23","doi-asserted-by":"crossref","unstructured":"Fan X, Deng Z, Wang K, Peng X, Qiao Y (2020) Learning discriminative representation for facial expression recognition from uncertainties. In: 2020 IEEE International Conference on Image Processing (ICIP), pp. 903\u2013907. IEEE, Virtual","DOI":"10.1109\/ICIP40778.2020.9190643"},{"key":"8040_CR24","doi-asserted-by":"crossref","unstructured":"Siqueira H, Magg S, Wermter S (2020) Efficient facial feature learning with wide ensemble-based convolutional neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 5800\u20135809. AAAI, New York","DOI":"10.1609\/aaai.v34i04.6037"},{"key":"8040_CR25","doi-asserted-by":"crossref","unstructured":"Liu P, Lin Y, Meng Z, Lu L, Deng W, Zhou JT, Yang Y (2021) Point adversarial self-mining: a simple method for facial expression recognition. IEEE T Cybern, pp 1\u201312","DOI":"10.1109\/TCYB.2021.3085744"},{"key":"8040_CR26","doi-asserted-by":"crossref","unstructured":"Ruan D, Yan Y, Lai S, Chai Z, Shen C, Wang H (2021) Feature decomposition and reconstruction learning for effective facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7660\u20137669. IEEE, Virtual","DOI":"10.1109\/CVPR46437.2021.00757"},{"key":"8040_CR27","doi-asserted-by":"crossref","unstructured":"Ma F, Sun B, Li S (2021) Facial expression recognition with visual transformers and attentional selective fusion. IEEE Trans Affect Comput","DOI":"10.1109\/TAFFC.2021.3122146"},{"key":"8040_CR28","doi-asserted-by":"crossref","unstructured":"Zhao Z, Liu Q, Zhou F (2021) Robust lightweight facial expression recognition network with label distribution training. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35(4):3510\u20133519. AAAI, Virtual","DOI":"10.1609\/aaai.v35i4.16465"},{"issue":"5","key":"8040_CR29","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/TIP.2018.2886767","volume":"28","author":"Y Li","year":"2018","unstructured":"Li Y, Zeng J, Shan S, Chen X (2018) Occlusion aware facial expression recognition using cnn with attention mechanism. IEEE Trans Image Process 28(5):2439\u20132450","journal-title":"IEEE Trans Image Process"},{"key":"8040_CR30","doi-asserted-by":"crossref","unstructured":"Albanie S, Nagrani A, Vedaldi A, Zisserman A (2018) Emotion recognition in speech using cross-modal transfer in the wild. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 292\u2013301. ACM, New York","DOI":"10.1145\/3240508.3240578"},{"key":"8040_CR31","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang K, Peng X, Yang J, Meng D, Qiao Y (2020) Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans Image Process 29:4057\u20134069","journal-title":"IEEE Trans Image Process"},{"issue":"2","key":"8040_CR32","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1109\/MMUL.2021.3076834","volume":"28","author":"H Xia","year":"2021","unstructured":"Xia H, Li C, Tan Y, Li L, Song S (2021) Destruction and reconstruction learning for facial expression recognition. IEEE Multimed 28(2):20\u201328","journal-title":"IEEE Multimed"},{"key":"8040_CR33","unstructured":"Li Y, Lu G, Li J, Zhang Z, Zhang D (2020) Facial expression recognition in the wild using multi-level features and attention mechanisms. IEEE Trans Affect Comput"},{"key":"8040_CR34","doi-asserted-by":"publisher","first-page":"6544","DOI":"10.1109\/TIP.2021.3093397","volume":"30","author":"Z Zhao","year":"2021","unstructured":"Zhao Z, Liu Q, Wang S (2021) Learning deep global multi-scale and local attention features for facial expression recognition in the wild. IEEE Trans Image Process 30:6544\u20136556","journal-title":"IEEE Trans Image Process"},{"key":"8040_CR35","unstructured":"M\u00fcller R, Kornblith S, Hinton GE (2019) When does label smoothing help? In: 33rd Conference on Neural Information Processing Systems (NIPS), vol. 32. MIT, Vancouver"},{"key":"8040_CR36","doi-asserted-by":"crossref","unstructured":"Goodfellow IJ, Erhan D, Carrier PL, Courville A, Mirza M, Hamner B, Cukierski W, Tang Y, Thaler D, Lee DH, et al. (2013) Challenges in representation learning: A report on three machine learning contests. In: International Conference on Neural Information Processing, pp. 117\u2013124. Springer, Daegu","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"8040_CR37","doi-asserted-by":"crossref","unstructured":"Guo Y, Zhang L, Hu Y, He X, Gao J (2016) Ms-celeb-1m: a dataset and benchmark for large-scale face recognition. In: European Conference on Computer Vision (ECCV), pp. 87\u2013102. Springer, Amsterdam","DOI":"10.1007\/978-3-319-46487-9_6"},{"key":"8040_CR38","doi-asserted-by":"crossref","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2017) Grad-cam: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 618\u2013626. IEEE, Venice","DOI":"10.1109\/ICCV.2017.74"},{"issue":"11","key":"8040_CR39","first-page":"2579","volume":"9","author":"L Van der Maaten","year":"2008","unstructured":"Van der Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9(11):2579\u20132605","journal-title":"J Mach Learn Res"},{"key":"8040_CR40","doi-asserted-by":"crossref","unstructured":"Huang C (2017) Combining convolutional neural networks for emotion recognition. In: 2017 IEEE MIT Undergraduate Research Technology Conference, pp. 1\u20134. IEEE, Massachusetts","DOI":"10.1109\/URTC.2017.8284175"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-08040-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-022-08040-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-08040-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,28]],"date-time":"2023-02-28T20:28:33Z","timestamp":1677616113000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-022-08040-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,11,22]]},"references-count":40,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2023,3]]}},"alternative-id":["8040"],"URL":"https:\/\/doi.org\/10.1007\/s00521-022-08040-4","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,11,22]]},"assertion":[{"value":"28 April 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 November 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 November 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We wish to submit a new manuscript entitled \u201cMPCSAN: Multi-head Parallel Channel-Spatial Attention Network for Facial Expression Recognition in the Wild\u201d for consideration in the Neural Computing and Applications Journal. We declare that this work is original and there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}