{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,23]],"date-time":"2025-11-23T12:21:38Z","timestamp":1763900498627,"version":"3.45.0"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2025,10,21]],"date-time":"2025-10-21T00:00:00Z","timestamp":1761004800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,21]],"date-time":"2025-10-21T00:00:00Z","timestamp":1761004800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s11760-025-04862-6","type":"journal-article","created":{"date-parts":[[2025,10,21]],"date-time":"2025-10-21T05:33:57Z","timestamp":1761024837000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A dual-interleaved feature differentiation network for mental health state recognition via facial expression analysis"],"prefix":"10.1007","volume":"19","author":[{"given":"Ying","family":"Wei","sequence":"first","affiliation":[]},{"given":"Yuetao","family":"Zheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,21]]},"reference":[{"key":"4862_CR1","doi-asserted-by":"crossref","unstructured":"Turcian, D., Stoicu-Tivadar, V.: Real-time detection of emotions based on facial expression for mental health. In: Telehealth Ecosystems in Practice, pp. 272\u2013276. IOS (2023)","DOI":"10.3233\/SHTI230795"},{"issue":"5","key":"4862_CR2","doi-asserted-by":"publisher","first-page":"394","DOI":"10.1007\/s11760-025-03984-1","volume":"19","author":"M Mohana","year":"2025","unstructured":"Mohana, M., Subashini, P., Ghinea, G.: XAI-DSCSA: explainable-AI-based deep semi-supervised convolutional sparse autoencoder for facial expression recognition. Signal. Image Video Process. 19(5), 394 (2025)","journal-title":"Signal. Image Video Process."},{"key":"4862_CR3","doi-asserted-by":"crossref","unstructured":"Hebri, D., Nuthakki, R., Digal, A.K., Venkatesan, K.G.S., Chawla, S., Reddy, R.: C.: Effective facial expression recognition system using machine learning. EAI Endorsed Trans. Internet Things, 10, (2024)","DOI":"10.4108\/eetiot.5362"},{"issue":"4","key":"4862_CR4","doi-asserted-by":"publisher","first-page":"325","DOI":"10.1007\/s11760-025-03889-z","volume":"19","author":"D Shao","year":"2025","unstructured":"Shao, D., Zhuang, L., Ma, L., Yi, S.: Expression recognition method based on feature redundancy optimization. Signal. Image Video Process. 19(4), 325 (2025)","journal-title":"Signal. Image Video Process."},{"issue":"17","key":"4862_CR5","doi-asserted-by":"publisher","first-page":"12717","DOI":"10.1007\/s00521-023-08372-9","volume":"35","author":"FM Talaat","year":"2023","unstructured":"Talaat, F.M.: Real-time facial emotion recognition system among children with autism based on deep learning and IoT. Neural Comput. Appl. 35(17), 12717\u201312728 (2023)","journal-title":"Neural Comput. Appl."},{"issue":"3","key":"4862_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11760-025-03822-4","volume":"19","author":"N Al-Garaawi","year":"2025","unstructured":"Al-Garaawi, N., Morris, T., Cootes, T.F.: Automatic facial expression localization and recognition across a large range of emotions. Signal. Image Video Process. 19(3), 1\u201311 (2025)","journal-title":"Signal. Image Video Process."},{"issue":"2","key":"4862_CR7","doi-asserted-by":"publisher","first-page":"199","DOI":"10.3390\/biomimetics8020199","volume":"8","author":"Z Wen","year":"2023","unstructured":"Wen, Z., Lin, W., Wang, T., Xu, G.: Distract your attention: Multi-head cross attention network for facial expression recognition. Biomimetics. 8(2), 199 (2023)","journal-title":"Biomimetics"},{"key":"4862_CR8","doi-asserted-by":"crossref","unstructured":"Awana, A., Singh, S.V., Mishra, A., Bhutani, V., Kumar, S.R., Shrivastava, P.: Live Emotion Detection Using Deepface. In 2023 6th International Conference on Contemporary Computing and Informatics (IC3I), 6, 581\u2013584, IEEE. (2023)","DOI":"10.1109\/IC3I59117.2023.10397747"},{"key":"4862_CR9","doi-asserted-by":"publisher","first-page":"100545","DOI":"10.1016\/j.cosrev.2023.100545","volume":"48","author":"SC Leong","year":"2023","unstructured":"Leong, S.C., Tang, Y.M., Lai, C.H., Lee, C.K.M.: Facial expression and body gesture emotion recognition: A systematic review on the use of visual data in affective computing. Comput. Sci. Rev. 48, 100545 (2023)","journal-title":"Comput. Sci. Rev."},{"issue":"9","key":"4862_CR10","doi-asserted-by":"publisher","first-page":"6695","DOI":"10.1007\/s00500-023-09477-y","volume":"28","author":"FM Talaat","year":"2024","unstructured":"Talaat, F.M., Ali, Z.H., Mostafa, R.R., El-Rashidy, N.: Real-time facial emotion recognition model based on kernel autoencoder and convolutional neural network for autism children. Soft. Comput. 28(9), 6695\u20136708 (2024)","journal-title":"Soft. Comput."},{"key":"4862_CR11","unstructured":"Alhussan, A.A., Talaat, F.M., El-kenawy, E.S.M., Abdelhamid, A.A., Ibrahim, A., Khafaga, D.S., Alnaggar, M., Computers: Mater. Continua, 76(1), (2023)"},{"key":"4862_CR12","unstructured":"Selvan, M.A.: Deep Learning Techniques for Comprehensive Emotion Recognition and Behavioral Regulation. (2024)"},{"issue":"31","key":"4862_CR13","doi-asserted-by":"publisher","first-page":"22935","DOI":"10.1007\/s00521-022-06913-2","volume":"35","author":"A Sharma","year":"2023","unstructured":"Sharma, A., Sharma, K., Kumar, A.: Real-time emotional health detection using fine-tuned transfer networks with multimodal fusion. Neural Comput. Appl. 35(31), 22935\u201322948 (2023)","journal-title":"Neural Comput. Appl."},{"key":"4862_CR14","doi-asserted-by":"crossref","unstructured":"Li, H., Niu, H., Zhu, Z., Zhao, F.: Cliper: A unified vision-language framework for in-the-wild facial expression recognition. In 2024 IEEE International Conference on Multimedia and Expo (ICME), 1\u20136, IEEE. (2024)","DOI":"10.1109\/ICME57554.2024.10687508"},{"issue":"12","key":"4862_CR15","doi-asserted-by":"publisher","first-page":"5475","DOI":"10.3390\/s23125475","volume":"23","author":"D Mamieva","year":"2023","unstructured":"Mamieva, D., Abdusalomov, A.B., Kutlimuratov, A., Muminov, B., Whangbo, T.K.: Multimodal emotion detection via attention-based fusion of extracted facial and speech features. Sensors. 23(12), 5475 (2023)","journal-title":"Sensors"},{"key":"4862_CR16","doi-asserted-by":"publisher","first-page":"33061","DOI":"10.1109\/ACCESS.2023.3263670","volume":"11","author":"S Wang","year":"2023","unstructured":"Wang, S., Qu, J., Zhang, Y., Zhang, Y.: Multimodal emotion recognition from EEG signals and facial expressions. IEEE Access. 11, 33061\u201333068 (2023)","journal-title":"IEEE Access."},{"key":"4862_CR17","doi-asserted-by":"crossref","unstructured":"Li, H., Wang, N., Yang, X., Wang, X., Gao, X.: Towards semi-supervised deep facial expression recognition with an adaptive confidence margin. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, 4166\u20134175, (2022)","DOI":"10.1109\/CVPR52688.2022.00413"},{"key":"4862_CR18","doi-asserted-by":"publisher","first-page":"2016","DOI":"10.1109\/TIP.2021.3049955","volume":"30","author":"H Li","year":"2021","unstructured":"Li, H., Wang, N., Ding, X., Yang, X., Gao, X.: Adaptively learning facial expression representation via Cf labels and distillation. IEEE Trans. Image Process. 30, 2016\u20132028 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"4862_CR19","doi-asserted-by":"crossref","unstructured":"Wang, H., Li, B., Wu, S., Shen, S., Liu, F., Ding, S., Zhou, A.: Rethinking the learning paradigm for dynamic facial expression recognition. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, 17958\u201317968, (2023)","DOI":"10.1109\/CVPR52729.2023.01722"},{"issue":"6","key":"4862_CR20","doi-asserted-by":"publisher","first-page":"15711","DOI":"10.1007\/s11042-023-16174-3","volume":"83","author":"G Meena","year":"2024","unstructured":"Meena, G., Mohbey, K.K., Indian, A., Khan, M.Z., Kumar, S.: Identifying emotions from facial expressions using a deep convolutional neural network-based approach. Multimedia Tools Appl. 83(6), 15711\u201315732 (2024)","journal-title":"Multimedia Tools Appl."},{"issue":"7","key":"4862_CR21","doi-asserted-by":"publisher","first-page":"78","DOI":"10.3390\/bdcc8070078","volume":"8","author":"D Sharma","year":"2024","unstructured":"Sharma, D., Singh, J., Sehra, S.S., Sehra, S.K.: Demystifying mental health by decoding facial action unit sequences. Big Data Cogn. Comput. 8(7), 78 (2024)","journal-title":"Big Data Cogn. Comput."},{"key":"4862_CR22","unstructured":"Parikh, A., Sadeghi, M., Eskofier, B.: Exploring Facial Biomarkers for Depression through Temporal Analysis of Action Units. arXiv preprint arXiv:2407.13753. (2024)"},{"issue":"2","key":"4862_CR23","doi-asserted-by":"publisher","first-page":"279","DOI":"10.26555\/jiteki.v10i2.28863","volume":"10","author":"F Fatimatuzzahra","year":"2024","unstructured":"Fatimatuzzahra, F., Lindawati, L., Soim, S.: Development of convolutional neural network models to improve facial expression recognition accuracy. Jurnal Ilmiah Teknik Elektro Komputer dan. Informatika. 10(2), 279\u2013289 (2024)","journal-title":"Jurnal Ilmiah Teknik Elektro Komputer dan. Informatika"},{"issue":"02","key":"4862_CR24","doi-asserted-by":"publisher","first-page":"2250075","DOI":"10.1142\/S0219649222500757","volume":"22","author":"A Jonnalagadda","year":"2023","unstructured":"Jonnalagadda, A., Rajvir, M., Singh, S., Chandramouliswaran, S., George, J., Kamalov, F.: An ensemble-based machine learning model for emotion and mental health detection. J. Inform. Knowl. Manage. 22(02), 2250075 (2023)","journal-title":"J. Inform. Knowl. Manage."},{"issue":"4","key":"4862_CR25","doi-asserted-by":"publisher","first-page":"1005","DOI":"10.1109\/TCE.2023.3263672","volume":"69","author":"X Ji","year":"2023","unstructured":"Ji, X., Dong, Z., Han, Y., Lai, C.S., Zhou, G., Qi, D.: EMSN: An energy-efficient memristive sequencer network for human emotion classification in mental health monitoring. IEEE Trans. Consum. Electron. 69(4), 1005\u20131016 (2023)","journal-title":"IEEE Trans. Consum. Electron."},{"issue":"2","key":"4862_CR26","doi-asserted-by":"publisher","first-page":"2214388","DOI":"10.1080\/20008066.2023.2214388","volume":"14","author":"LL Hautle","year":"2023","unstructured":"Hautle, L.L., Jellestad, L., Schenkel, S., Wingenbach, T.S., Peyk, P., Schnyder, U., Weilenmann, S., Pfaltz, M.C.: Adults with a history of childhood maltreatment with and without mental disorders show alterations in the recognition of facial expressions. Eur. J. Psychotraumatology. 14(2), 2214388 (2023)","journal-title":"Eur. J. Psychotraumatology"},{"key":"4862_CR27","doi-asserted-by":"publisher","first-page":"107923","DOI":"10.1016\/j.cmpb.2023.107923","volume":"243","author":"X Li","year":"2024","unstructured":"Li, X., Yi, X., Ye, J., Zheng, Y., Wang, Q.: SFTNet: A microexpression-based method for depression detection. Comput. Methods Programs Biomed. 243, 107923 (2024)","journal-title":"Comput. Methods Programs Biomed."},{"key":"4862_CR28","doi-asserted-by":"publisher","first-page":"4637","DOI":"10.1109\/TIP.2022.3186536","volume":"31","author":"H Li","year":"2022","unstructured":"Li, H., Wang, N., Yang, X., Gao, X.: Crs-cont: A well-trained general encoder for facial expression analysis. IEEE Trans. Image Process. 31, 4637\u20134650 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"4862_CR29","doi-asserted-by":"publisher","first-page":"119640","DOI":"10.1016\/j.ins.2023.119640","volume":"649","author":"F Jiang","year":"2023","unstructured":"Jiang, F., et al.: Face2nodes: Learning facial expression representations with relation-aware dynamic graph Convolution networks. Inf. Sci. 649, 119640 (2023)","journal-title":"Inf. Sci."},{"key":"4862_CR30","doi-asserted-by":"crossref","unstructured":"Huang, C., et al.: Modeling Fine-Grained relations in dynamic Space-Time graphs for Video-Based facial expression recognition. IEEE Trans. Affect. Comput. (2025)","DOI":"10.1109\/TAFFC.2025.3530973"},{"key":"4862_CR31","doi-asserted-by":"crossref","unstructured":"Li, H., Yang, Y., Zhong, H., Zhao, J.: Predicting Suicide Risk Using Facial Action Units: A Machine Learning Approach to Objective and Large-Scale Screening. (2024)","DOI":"10.21203\/rs.3.rs-4185615\/v1"},{"key":"4862_CR32","doi-asserted-by":"crossref","unstructured":"Mineur, L., Heide, M., Eickhoff, S., Avram, M., Franzen, L., Buschmann, F., Schr\u00f6pfer, F., Rogg, H.V., Andreou, C., Br\u00fcgge, N.S., Handels, H., Borgwardt, S., Korda, A.: Analysis of facial expressions recorded from patients during psychiatric interviews. (2024). medRxiv, 2024-09.","DOI":"10.1101\/2024.09.19.24313994"},{"issue":"5","key":"4862_CR33","first-page":"172","volume":"17","author":"S Sugiyanto","year":"2024","unstructured":"Sugiyanto, S., Purnama, I.K.E., Yuniarno, E.M., Anggraeni, W., Purnomo, M.H.: Depression classification based on facial action unit intensity features using CNN-Poolingless framework. Int. J. Intell. Eng. Syst. 17(5), 172\u2013187 (2024)","journal-title":"Int. J. Intell. Eng. Syst."},{"issue":"4","key":"4862_CR34","doi-asserted-by":"publisher","first-page":"1967","DOI":"10.1007\/s00530-023-01080-3","volume":"29","author":"G Wang","year":"2023","unstructured":"Wang, G., Huang, S., Tao, Z.: Shallow multi-branch attention convolutional neural network for micro-expression recognition. Multimedia Syst. 29(4), 1967\u20131980 (2023)","journal-title":"Multimedia Syst."},{"issue":"3","key":"4862_CR35","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1007\/s00530-024-01352-6","volume":"30","author":"G Wang","year":"2024","unstructured":"Wang, G., Huang, S.: Dual-stream network with cross-layer attention and similarity constraint for micro-expression recognition. Multimedia Syst. 30(3), 147 (2024)","journal-title":"Multimedia Syst."},{"key":"4862_CR36","doi-asserted-by":"crossref","unstructured":"Wang, G., Zou, Y., He, S., Wang, Y., Dai, R.: Anomaly Detection and Localization Via Reverse Distillation with Latent Anomaly Suppression. IEEE Transactions on Circuits and Systems for Video Technology (2025)","DOI":"10.1109\/TCSVT.2025.3562258"},{"issue":"6","key":"4862_CR37","doi-asserted-by":"publisher","first-page":"176","DOI":"10.1007\/s10462-025-11159-0","volume":"58","author":"G Liu","year":"2025","unstructured":"Liu, G., Huang, S., Wang, G., Li, M.: Emrnet: Enhanced micro-expression recognition network with attention and distance correlation. Artif. Intell. Rev. 58(6), 176 (2025)","journal-title":"Artif. Intell. Rev."},{"issue":"1","key":"4862_CR38","doi-asserted-by":"publisher","first-page":"14429","DOI":"10.1038\/s41598-024-65276-x","volume":"14","author":"ES Agung","year":"2024","unstructured":"Agung, E.S., Rifai, A.P., Wijayanto, T.: Image-based facial emotion recognition using convolutional neural network on emognition dataset. Sci. Rep. 14(1), 14429 (2024)","journal-title":"Sci. Rep."},{"key":"4862_CR39","doi-asserted-by":"publisher","first-page":"41557","DOI":"10.1109\/ACCESS.2024.3377235","volume":"12","author":"JH Kim","year":"2024","unstructured":"Kim, J.H., Poulose, A., Han, D.S.: Cvgg-19: Customized visual geometry group deep learning architecture for facial emotion recognition. IEEE Access. 12, 41557\u201341578 (2024)","journal-title":"IEEE Access."},{"key":"4862_CR40","doi-asserted-by":"crossref","unstructured":"Roy, A.K., Kathania, H.K., Sharma, A., Dey, A., Ansari, M.S.A.: ResEmoteNet: Bridging accuracy and loss reduction in facial emotion recognition. IEEE. Signal. Process. Lett., (2024)","DOI":"10.36227\/techrxiv.172651476.62062165\/v1"},{"issue":"4","key":"4862_CR41","doi-asserted-by":"publisher","first-page":"3723","DOI":"10.1080\/03772063.2023.2202158","volume":"70","author":"S Kaur","year":"2024","unstructured":"Kaur, S., Kulkarni, N.: FERFM: An enhanced facial emotion recognition system using Fine-tuned MobileNetV2 architecture. IETE J. Res. 70(4), 3723\u20133737 (2024)","journal-title":"IETE J. Res."},{"key":"4862_CR42","doi-asserted-by":"crossref","unstructured":"Kumar, R.J.R., Leena, M.A., Saumiya, S., Gnanaprakash, V., Febin, U., Stanley, B.F.: Attention-Guided Facial Emotion Recognition with Swin Transformer Encoder Decoder Model. In 2024 International Conference on Advancement in Renewable Energy and Intelligent Systems (AREIS), 1\u20136, IEEE. (2024)","DOI":"10.1109\/AREIS62559.2024.10893667"},{"key":"4862_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Liu, B., Zhou, J., Wang, H., Liu, X., Lin, B., Chen, T.: Masked facial expression recognition based on Temporal overlap module and action unit graph convolutional network. J. Vis. Commun. Image Represent., 104398, (2025)","DOI":"10.1016\/j.jvcir.2025.104398"},{"key":"4862_CR44","doi-asserted-by":"crossref","unstructured":"Savchenko, A.V., Sidorova, A.P.: EmotiEffNet and Temporal Convolutional Networks in Video-based Facial Expression Recognition and Action Unit Detection. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 4849\u20134859, (2024)","DOI":"10.1109\/CVPRW63382.2024.00488"},{"key":"4862_CR45","doi-asserted-by":"publisher","first-page":"2067","DOI":"10.1109\/TIP.2025.3546479","volume":"34","author":"Y Li","year":"2025","unstructured":"Li, Y., Liu, M., Cui, Z., Ding, Y., Zong, Y., Zheng, W., Shan, S., Guan, C.: Decoupled doubly contrastive learning for cross domain facial action unit detection. IEEE Trans. Image Process. 34, 2067\u20132080 (2025)","journal-title":"IEEE Trans. Image Process."},{"key":"4862_CR46","unstructured":"Tan, P.S., Rajanala, S., Pal, A., Phan, R.C.W., Ong, H.F.: Causal-Ex: Causal Graph-based micro and macro expression spotting. (2025). arXiv preprint arXiv:2503.09098.2"},{"key":"4862_CR47","doi-asserted-by":"crossref","unstructured":"Flores, R., Tlachac, M.L., Shrestha, A., Rundensteiner, E.A.: WavFace: A multimodal Transformer-based model for depression screening. IEEE J. Biomedical Health Inf., (2025)","DOI":"10.1109\/JBHI.2025.3529348"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04862-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04862-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04862-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,23]],"date-time":"2025-11-23T12:13:54Z","timestamp":1763900034000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04862-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,21]]},"references-count":47,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4862"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04862-6","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,10,21]]},"assertion":[{"value":"27 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 September 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 October 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1305"}}