{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,6]],"date-time":"2025-12-06T08:03:49Z","timestamp":1765008229584,"version":"3.46.0"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["61977027","61977027","61977027"],"award-info":[{"award-number":["61977027","61977027","61977027"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100002858","name":"the China Postdoctoral Science Foundation","doi-asserted-by":"crossref","award":["2023M742718","2023M742718","2023M742718"],"award-info":[{"award-number":["2023M742718","2023M742718","2023M742718"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Pattern Anal Applic"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s10044-025-01499-6","type":"journal-article","created":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T14:36:44Z","timestamp":1757342204000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dual-stream network with coordinate attention for multi-view micro-expression recognition using 3D face reconstruction"],"prefix":"10.1007","volume":"28","author":[{"given":"Pianpian","family":"Ma","sequence":"first","affiliation":[]},{"given":"Jingying","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xiaodi","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,8]]},"reference":[{"issue":"2","key":"1499_CR1","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1037\/h0030377","volume":"17","author":"P Ekman","year":"1971","unstructured":"Ekman P, Friesen WV (1971) Constants across cultures in the face and emotion. J Pers Soc Psychol 17(2):124","journal-title":"J Pers Soc Psychol"},{"key":"1499_CR2","doi-asserted-by":"crossref","unstructured":"Zhang J, Dong Z, Wang S-J (2023) Micro expression recognition by machine learning based profit function analysis in intelligent marketing of financial industry. In: 2023 IEEE 6th Eurasian Conference on Educational Innovation (ECEI), pp. 188\u2013192. IEEE","DOI":"10.1109\/ECEI57668.2023.10105404"},{"issue":"10","key":"1499_CR3","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1109\/MCOM.001.2300003","volume":"61","author":"X Chen","year":"2023","unstructured":"Chen X, Luo T (2023) Catching elusive depression via facial micro-expression recognition. IEEE Commun Mag 61(10):30\u201336","journal-title":"IEEE Commun Mag"},{"issue":"1","key":"1499_CR4","doi-asserted-by":"publisher","first-page":"469","DOI":"10.1111\/bjet.12597","volume":"50","author":"M-H Chiu","year":"2019","unstructured":"Chiu M-H, Liaw HL, Yu Y-R, Chou C-C (2019) Facial micro-expression states as an indicator for conceptual change in students\u2019 understanding of air pressure and boiling points. Br J Edu Technol 50(1):469\u2013480","journal-title":"Br J Edu Technol"},{"issue":"2","key":"1499_CR5","first-page":"5","volume":"1","author":"P Ekman","year":"2009","unstructured":"Ekman P (2009) Lie catching and microexpressions. Philos Decept 1(2):5","journal-title":"Philos Decept"},{"issue":"4","key":"1499_CR6","doi-asserted-by":"publisher","first-page":"384","DOI":"10.1037\/0003-066X.48.4.384","volume":"48","author":"P Ekman","year":"1993","unstructured":"Ekman P (1993) Facial expression and emotion. Am Psychol 48(4):384","journal-title":"Am Psychol"},{"key":"1499_CR7","doi-asserted-by":"crossref","unstructured":"Van\u00a0Quang N, Chun J, Tokuyama T (2019) Capsulenet for micro-expression recognition. In: 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1\u20137. IEEE","DOI":"10.1109\/FG.2019.8756544"},{"issue":"2","key":"1499_CR8","doi-asserted-by":"publisher","first-page":"1037","DOI":"10.1109\/TAFFC.2020.2986962","volume":"13","author":"B Sun","year":"2020","unstructured":"Sun B, Cao S, Li D, He J, Yu L (2020) Dynamic micro-expression recognition using knowledge distillation. IEEE Trans Affect Comput 13(2):1037\u20131043","journal-title":"IEEE Trans Affect Comput"},{"key":"1499_CR9","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1109\/TIP.2020.3035042","volume":"30","author":"Y Li","year":"2020","unstructured":"Li Y, Huang X, Zhao G (2020) Joint local and global information learning with single apex frame detection for micro-expression recognition. IEEE Trans Image Process 30:249\u2013263","journal-title":"IEEE Trans Image Process"},{"issue":"3","key":"1499_CR10","doi-asserted-by":"publisher","first-page":"1527","DOI":"10.1007\/s00371-024-03443-x","volume":"41","author":"J Yang","year":"2025","unstructured":"Yang J, Wu Z, Wu R (2025) Micro-expression recognition based on contextual transformer networks. Vis Comput 41(3):1527\u20131541","journal-title":"Vis Comput"},{"issue":"4","key":"1499_CR11","doi-asserted-by":"publisher","first-page":"3031","DOI":"10.1109\/TAFFC.2022.3182342","volume":"14","author":"X Li","year":"2022","unstructured":"Li X, Cheng S, Li Y, Behzad M, Shen J, Zafeiriou S, Pantic M, Zhao G (2022) 4dme: a spontaneous 4d micro-expression dataset with multimodalities. IEEE Trans Affect Comput 14(4):3031\u20133047","journal-title":"IEEE Trans Affect Comput"},{"issue":"16","key":"1499_CR12","doi-asserted-by":"publisher","first-page":"2181","DOI":"10.1016\/j.patrec.2012.07.015","volume":"33","author":"X Huang","year":"2012","unstructured":"Huang X, Zhao G, Zheng W, Pietik\u00e4inen M (2012) Towards a dynamic expression recognition system under facial occlusion. Pattern Recogn Lett 33(16):2181\u20132191","journal-title":"Pattern Recogn Lett"},{"issue":"1","key":"1499_CR13","doi-asserted-by":"publisher","first-page":"86041","DOI":"10.1371\/journal.pone.0086041","volume":"9","author":"W-J Yan","year":"2014","unstructured":"Yan W-J, Li X, Wang S-J, Zhao G, Liu Y-J, Chen Y-H, Fu X (2014) Casme ii: an improved spontaneous micro-expression database and the baseline evaluation. PLoS One 9(1):86041","journal-title":"PLoS One"},{"key":"1499_CR14","doi-asserted-by":"crossref","unstructured":"Li X, Pfister T, Huang X, Zhao G, Pietik\u00e4inen M (2013) A spontaneous micro-expression database: inducement, collection and baseline. In: 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (fg), pp. 1\u20136. IEEE","DOI":"10.1109\/FG.2013.6553717"},{"issue":"1","key":"1499_CR15","doi-asserted-by":"publisher","first-page":"116","DOI":"10.1109\/TAFFC.2016.2573832","volume":"9","author":"AK Davison","year":"2016","unstructured":"Davison AK, Lansley C, Costen N, Tan K, Yap MH (2016) Samm: a spontaneous micro-facial movement dataset. IEEE Trans Affect Comput 9(1):116\u2013129","journal-title":"IEEE Trans Affect Comput"},{"key":"1499_CR16","doi-asserted-by":"crossref","unstructured":"See J, Yap MH, Li J, Hong X, Wang S-J (2019) Megc 2019\u2013the second facial micro-expressions grand challenge. In: 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1\u20135. IEEE","DOI":"10.1109\/FG.2019.8756611"},{"issue":"6","key":"1499_CR17","doi-asserted-by":"publisher","first-page":"915","DOI":"10.1109\/TPAMI.2007.1110","volume":"29","author":"G Zhao","year":"2007","unstructured":"Zhao G, Pietikainen M (2007) Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans Pattern Anal Mach Intell 29(6):915\u2013928","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1499_CR18","doi-asserted-by":"crossref","unstructured":"Wang Y, See J, Phan RC-W, Oh Y-H (2014) Lbp with six intersection points: Reducing redundant information in lbp-top for micro-expression recognition. In: Asian Conference on Computer Vision, pp. 525\u2013537. Springer","DOI":"10.1007\/978-3-319-16865-4_34"},{"issue":"12","key":"1499_CR19","doi-asserted-by":"publisher","first-page":"130","DOI":"10.3390\/jimaging6120130","volume":"6","author":"AM Buhari","year":"2020","unstructured":"Buhari AM, Ooi C-P, Baskaran VM, Phan RC, Wong K, Tan W-H (2020) Facs-based graph features for real-time micro-expression recognition. J Imaging 6(12):130","journal-title":"J Imaging"},{"key":"1499_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.129135","volume":"619","author":"L Zhang","year":"2025","unstructured":"Zhang L, Zhang Y, Sun X, Tang W, Wang X, Li Z (2025) Micro-expression recognition based on direct learning of graph structure. Neurocomputing 619:129135","journal-title":"Neurocomputing"},{"issue":"3","key":"1499_CR21","doi-asserted-by":"publisher","first-page":"194","DOI":"10.1038\/35058500","volume":"2","author":"L Itti","year":"2001","unstructured":"Itti L, Koch C (2001) Computational modelling of visual attention. Nat Rev Neurosci 2(3):194\u2013203","journal-title":"Nat Rev Neurosci"},{"key":"1499_CR22","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1016\/j.neucom.2020.06.005","volume":"410","author":"C Wang","year":"2020","unstructured":"Wang C, Peng M, Bi T, Chen T (2020) Micro-attention for micro-expression recognition. Neurocomputing 410:354\u2013362","journal-title":"Neurocomputing"},{"key":"1499_CR23","doi-asserted-by":"crossref","unstructured":"Zhou L, Mao Q, Xue L (2019) Cross-database micro-expression recognition: a style aggregated and attention transfer approach. In: 2019 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 102\u2013107 (2019). IEEE","DOI":"10.1109\/ICMEW.2019.00025"},{"key":"1499_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2024.104383","volume":"107","author":"X Ding","year":"2025","unstructured":"Ding X, Li Y, Wu Y, Guo W (2025) Enhancement-suppression driven lightweight fine-grained micro-expression recognition. J Vis Commun Image Represent 107:104383","journal-title":"J Vis Commun Image Represent"},{"key":"1499_CR25","doi-asserted-by":"crossref","unstructured":"Peng M, Wang C, Bi T, Shi Y, Zhou X, Chen T (2019) A novel apex-time network for cross-dataset micro-expression recognition. In: 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 1\u20136. IEEE","DOI":"10.1109\/ACII.2019.8925525"},{"issue":"3","key":"1499_CR26","doi-asserted-by":"publisher","first-page":"1271","DOI":"10.1007\/s00521-024-10374-0","volume":"37","author":"N Matharaarachchi","year":"2025","unstructured":"Matharaarachchi N, Fermi Pasha M (2025) A dual stream spatio-temporal deep network for micro-expression recognition using upper facial features. Neural Comput Appl 37(3):1271\u20131287","journal-title":"Neural Comput Appl"},{"issue":"9","key":"1499_CR27","doi-asserted-by":"publisher","first-page":"1246","DOI":"10.3390\/e25091246","volume":"25","author":"H Yang","year":"2023","unstructured":"Yang H, Xie L, Pan H, Li C, Wang Z, Zhong J (2023) Multimodal attention dynamic fusion network for facial micro-expression recognition. Entropy 25(9):1246","journal-title":"Entropy"},{"key":"1499_CR28","doi-asserted-by":"crossref","unstructured":"Koujan MR, Alharbawee L, Giannakakis G, Pugeault N, Roussos A (2020) Real-time facial expression recognition \u201cin the wild\u201d by disentangling 3d expression from identity. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), pp. 24\u201331. IEEE","DOI":"10.1109\/FG47880.2020.00084"},{"key":"1499_CR29","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2019.10.003","volume":"92","author":"TS Ly","year":"2019","unstructured":"Ly TS, Do N-T, Kim S-H, Yang H-J, Lee G-S (2019) A novel 2d and 3d multimodal approach for in-the-wild facial expression recognition. Image Vis Comput 92:103817","journal-title":"Image Vis Comput"},{"key":"1499_CR30","first-page":"2","volume":"1","author":"AT Tran","year":"2018","unstructured":"Tran AT, Hassner T, Masi I, Paz E, Nirkin Y, Medioni GG et al (2018) Extreme 3d face reconstruction: seeing through occlusions. CVPR 1:2","journal-title":"CVPR"},{"issue":"4","key":"1499_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459936","volume":"40","author":"Y Feng","year":"2021","unstructured":"Feng Y, Feng H, Black MJ, Bolkart T (2021) Learning an animatable detailed 3d face model from in-the-wild images. ACM Trans Graph (ToG) 40(4):1\u201313","journal-title":"ACM Trans Graph (ToG)"},{"key":"1499_CR32","doi-asserted-by":"crossref","unstructured":"Hou Q, Zhou D, Feng J (2021) Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13713\u201313722","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"1499_CR33","unstructured":"Hinton GE, Sabour S, Frosst N (2018) Matrix capsules with em routing. In: International Conference on learning representations"},{"key":"1499_CR34","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","volume":"128","author":"RR Selvaraju","year":"2020","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2020) Grad-cam: visual explanations from deep networks via gradient-based localization. Int J Comput Vis 128:336\u2013359","journal-title":"Int J Comput Vis"}],"container-title":["Pattern Analysis and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-025-01499-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10044-025-01499-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-025-01499-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,6]],"date-time":"2025-12-06T07:59:40Z","timestamp":1765007980000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10044-025-01499-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,8]]},"references-count":34,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["1499"],"URL":"https:\/\/doi.org\/10.1007\/s10044-025-01499-6","relation":{},"ISSN":["1433-7541","1433-755X"],"issn-type":[{"type":"print","value":"1433-7541"},{"type":"electronic","value":"1433-755X"}],"subject":[],"published":{"date-parts":[[2025,9,8]]},"assertion":[{"value":"10 January 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"167"}}