{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,21]],"date-time":"2026-04-21T06:41:19Z","timestamp":1776753679745,"version":"3.51.2"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T00:00:00Z","timestamp":1751500800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T00:00:00Z","timestamp":1751500800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s40747-025-01931-8","type":"journal-article","created":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T05:02:58Z","timestamp":1751518978000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":14,"title":["EMVAS: end-to-end multimodal emotion visualization analysis system"],"prefix":"10.1007","volume":"11","author":[{"given":"Xianxun","family":"Zhu","sequence":"first","affiliation":[]},{"given":"Heyang","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Erik","family":"Cambria","sequence":"additional","affiliation":[]},{"given":"Yao","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Ju","sequence":"additional","affiliation":[]},{"given":"Haochen","family":"Yuan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7974-9510","authenticated-orcid":false,"given":"Rui","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,3]]},"reference":[{"key":"1931_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2024.108564","volume":"260","author":"X Zhu","year":"2025","unstructured":"Zhu X, Liu Z, Cambria E, Yu X, Fan X, Chen H, Wang R (2025) A client-server based recognition system: non-contact single\/multiple emotional and behavioral state assessment methods. Comput Methods Programs Biomed 260:108564","journal-title":"Comput Methods Programs Biomed"},{"issue":"4","key":"1931_CR2","doi-asserted-by":"publisher","first-page":"4069","DOI":"10.1007\/s10639-022-11370-4","volume":"28","author":"I Lasri","year":"2023","unstructured":"Lasri I, Riadsolh A, Elbelkacemi M (2023) Facial emotion recognition of deaf and hard-of-hearing students for engagement detection using deep learning. Educ Inf Technol 28(4):4069\u20134092","journal-title":"Educ Inf Technol"},{"key":"1931_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.114900","volume":"177","author":"A Saibene","year":"2021","unstructured":"Saibene A, Assale M, Giltri M (2021) Expert systems: definitions, advantages and issues in medical field applications. Expert Syst Appl 177:114900","journal-title":"Expert Syst Appl"},{"key":"1931_CR4","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1016\/j.inffus.2017.02.003","volume":"37","author":"S Poria","year":"2017","unstructured":"Poria S, Cambria E, Bajpai R, Hussain A (2017) A review of affective computing: from unimodal analysis to multimodal fusion. Inf Fusion 37:98\u2013125","journal-title":"Inf Fusion"},{"issue":"1","key":"1931_CR5","doi-asserted-by":"publisher","first-page":"68","DOI":"10.1109\/TCE.2021.3056421","volume":"67","author":"R Chatterjee","year":"2021","unstructured":"Chatterjee R, Mazumdar S, Sherratt RS, Halder R, Maitra T, Giri D (2021) Real-time speech emotion analysis for smart home assistants. IEEE Trans Consum Electron 67(1):68\u201376","journal-title":"IEEE Trans Consum Electron"},{"issue":"18","key":"1931_CR6","doi-asserted-by":"publisher","first-page":"56039","DOI":"10.1007\/s11042-023-17347-w","volume":"83","author":"X Zhu","year":"2024","unstructured":"Zhu X, Huang Y, Wang X, Wang R (2024) Emotion recognition based on brain-like multimodal hierarchical perception. Multimed Tools Appl 83(18):56039\u201356057","journal-title":"Multimed Tools Appl"},{"key":"1931_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110525","volume":"270","author":"M Khan","year":"2023","unstructured":"Khan M, El Saddik A, Alotaibi FS, Pham NT (2023) AAD-Net: advanced end-to-end signal processing system for human emotion detection & recognition using attention-based deep echo state network. Knowl Based Syst 270:110525","journal-title":"Knowl Based Syst"},{"issue":"2","key":"1931_CR8","doi-asserted-by":"publisher","first-page":"377","DOI":"10.1007\/s11760-021-01943-0","volume":"16","author":"I Chaturvedi","year":"2022","unstructured":"Chaturvedi I, Chen Q, Cambria E, McConnell D (2022) Landmark calibration for facial expressions and fish classification. SIViP 16(2):377\u2013384","journal-title":"SIViP"},{"key":"1931_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.114683","volume":"173","author":"D Li","year":"2021","unstructured":"Li D, Liu J, Yang Z, Sun L, Wang Z (2021) Speech emotion recognition using recurrent neural networks with directional self-attention. Expert Syst Appl 173:114683","journal-title":"Expert Syst Appl"},{"issue":"4","key":"1931_CR10","doi-asserted-by":"publisher","first-page":"1504","DOI":"10.1007\/s12559-024-10287-z","volume":"16","author":"X Zhu","year":"2024","unstructured":"Zhu X, Guo C, Feng H, Huang Y, Feng Y, Wang X, Wang R (2024) A review of key technologies for emotion analysis using multimodal information. Cogn Comput 16(4):1504\u20131530","journal-title":"Cogn Comput"},{"issue":"4","key":"1931_CR11","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1007\/s13735-024-00347-3","volume":"13","author":"R Wang","year":"2024","unstructured":"Wang R, Zhu J, Wang S, Wang T, Huang J, Zhu X (2024) Multi-modal emotion recognition using tensor decomposition fusion and self-supervised multi-tasking. Int J Multimed Inf Retrieval 13(4):39","journal-title":"Int J Multimed Inf Retrieval"},{"key":"1931_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102306","volume":"106","author":"C Fan","year":"2024","unstructured":"Fan C, Lin J, Mao R, Cambria E (2024) Fusing pairwise modalities for emotion recognition in conversations. Inf Fusion 106:102306","journal-title":"Inf Fusion"},{"key":"1931_CR13","doi-asserted-by":"crossref","unstructured":"Williams J, Kleinegesse S, Comanescu R, Radu O (2018) Recognizing emotions in video using multimodal DNN feature fusion. In: Proceedings of the challenge-HML conference, pp 11\u201319","DOI":"10.18653\/v1\/W18-3302"},{"key":"1931_CR14","doi-asserted-by":"crossref","unstructured":"Yu W, Xu H, Meng F, Zhu Y, Ma Y, Wu J, Zou J, Yang K (2020) CH-SIMS: a Chinese multimodal sentiment analysis dataset with fine-grained annotation of modality. In: Proceedings of the ACL - annual meeting association computational linguistics conference, pp 3718\u20133727","DOI":"10.18653\/v1\/2020.acl-main.343"},{"key":"1931_CR15","doi-asserted-by":"crossref","unstructured":"Zadeh A, Chen M, Poria S, et al. (2017) Tensor fusion network for multimodal sentiment analysis. arXiv preprint arXiv:1707.07250","DOI":"10.18653\/v1\/D17-1115"},{"key":"1931_CR16","doi-asserted-by":"crossref","unstructured":"Liu Z, Shen Y, Lakshminarasimhan VB, Liang PP, Zadeh A, Morency L (2018) Efficient low-rank multimodal fusion with modality-specific factors. In: Proceedings of the ACL - annual meeting of the association for computational linguistics conference, pp 2247\u20132256","DOI":"10.18653\/v1\/P18-1209"},{"issue":"9","key":"1931_CR17","doi-asserted-by":"publisher","first-page":"5318","DOI":"10.1109\/TCSVT.2023.3247822","volume":"33","author":"M Hou","year":"2023","unstructured":"Hou M, Zhang Z, Liu C, Lu G (2023) Semantic alignment network for multi-modal emotion recognition. IEEE Trans Circuits Syst Video Technol 33(9):5318\u20135329","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"1931_CR18","doi-asserted-by":"crossref","unstructured":"Liang B, Gui L, He Y, Cambria E, Xu R (2024) Fusion and discrimination: a multimodal graph contrastive learning framework for multimodal sarcasm detection. IEEE Trans Affect Comput","DOI":"10.1109\/TAFFC.2024.3380375"},{"issue":"1","key":"1931_CR19","doi-asserted-by":"publisher","first-page":"320","DOI":"10.1109\/TAFFC.2020.3000510","volume":"13","author":"S Mai","year":"2020","unstructured":"Mai S, Hu H, Xu J, Xing S (2020) Multi-fusion residual memory network for multimodal human sentiment comprehension. IEEE Trans Affect Comput 13(1):320\u2013334","journal-title":"IEEE Trans Affect Comput"},{"key":"1931_CR20","unstructured":"Amodei D, Ananthanarayanan S, Anubhai R, Bai J, Battenberg E, Case C, Zhu Z (2016) Deep audio 2: end-to-end audio recognition in English and Mandarin. In: International conference on machine learning. PMLR, pp 173\u2013182"},{"key":"1931_CR21","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding, arXiv preprint arXiv:1810.04805"},{"issue":"1","key":"1931_CR22","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1146\/annurev.ps.30.020179.002523","volume":"30","author":"P Ekman","year":"1979","unstructured":"Ekman P, Oster H (1979) Facial expressions of emotion. Annu Rev Psychol 30(1):527\u2013554","journal-title":"Annu Rev Psychol"},{"key":"1931_CR23","volume-title":"What the face reveals: basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS)","year":"1997","unstructured":"Ekman P, Rosenberg EL (eds) (1997) What the face reveals: basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS). Oxford University Press, Oxford"},{"key":"1931_CR24","doi-asserted-by":"publisher","first-page":"307","DOI":"10.1016\/j.aej.2022.08.003","volume":"63","author":"L Yan","year":"2023","unstructured":"Yan L, Shi Y, Wei M, Wu Y (2023) Multi-feature fusing local directional ternary pattern for facial expressions signal recognition based on video communication system. Alex Eng J 63:307\u2013320","journal-title":"Alex Eng J"},{"key":"1931_CR25","doi-asserted-by":"crossref","unstructured":"Guo X, Zhang Y, Lu S, Lu Z (2023) Facial expression recognition: a review. In: Multimedia tools and applications, pp 1\u201347","DOI":"10.1007\/s11042-023-15982-x"},{"key":"1931_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.apacoust.2022.109178","volume":"202","author":"ZT Liu","year":"2023","unstructured":"Liu ZT, Han MT, Wu BH, Rehman A (2023) audio emotion recognition based on convolutional neural network with attention-based bidirectional long short-term memory network and multi-task learning. Appl Acoust 202:109178","journal-title":"Appl Acoust"},{"key":"1931_CR27","doi-asserted-by":"crossref","unstructured":"Mirsamadi S, Barsoum E, Zhang C (2017) Automatic audio emotion recognition using recurrent neural networks with local attention. In: 2017 IEEE international conference on acoustics, audio and signal processing (ICASSP), pp 2227\u20132231","DOI":"10.1109\/ICASSP.2017.7952552"},{"key":"1931_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.124104","volume":"252","author":"T Gu","year":"2024","unstructured":"Gu T, He Z, Zhao H, Li M, Ying D (2024) Aspect-based sentiment analysis with multi-granularity information mining and sentiment hint. Expert Syst Appl 252:124104","journal-title":"Expert Syst Appl"},{"key":"1931_CR29","doi-asserted-by":"crossref","unstructured":"McFee B, Raffel C, Liang D, Ellis DP, McVicar M, Battenberg E, Nieto O (2015) librosa: audio and music signal analysis in python. In: SciPy, pp 18\u201324","DOI":"10.25080\/Majora-7b98e3ed-003"},{"key":"1931_CR30","unstructured":"Culjak I, Abram D, Pribanic T, Dzapo H, Cifrek M (2012) A brief introduction to OpenCV. In: 2012 Proceedings of the 35th international convention MIPRO, pp 1725\u20131730"},{"key":"1931_CR31","unstructured":"Wang D, Zhang X (2015) THCHS-30: A free Chinese audio corpus, arXiv preprint arXiv:1512.01882"},{"key":"1931_CR32","doi-asserted-by":"crossref","unstructured":"Yu W, Xu H, Meng F, Zhu Y, Ma Y, Wu J, Yang K, Zou J (2020) CH-SIMS: a Chinese multimodal sentiment analysis dataset with fine-grained annotation of modality. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 3718\u20133727","DOI":"10.18653\/v1\/2020.acl-main.343"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01931-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-01931-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01931-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T00:38:57Z","timestamp":1757205537000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-01931-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,3]]},"references-count":32,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["1931"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-01931-8","relation":{"is-referenced-by":[{"id-type":"doi","id":"10.1038\/s41598-025-32521-w","asserted-by":"object"}]},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,3]]},"assertion":[{"value":"14 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 April 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"There is no conflict of interest in our work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"374"}}