{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,29]],"date-time":"2026-03-29T21:07:10Z","timestamp":1774818430609,"version":"3.50.1"},"reference-count":277,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,8,13]],"date-time":"2025-08-13T00:00:00Z","timestamp":1755043200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,8,13]],"date-time":"2025-08-13T00:00:00Z","timestamp":1755043200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Artif Intell Rev"],"DOI":"10.1007\/s10462-025-11271-1","type":"journal-article","created":{"date-parts":[[2025,8,12]],"date-time":"2025-08-12T23:42:32Z","timestamp":1755042152000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["A review and critical analysis of multimodal datasets for emotional AI"],"prefix":"10.1007","volume":"58","author":[{"given":"Sadam","family":"Al-Azani","sequence":"first","affiliation":[]},{"given":"El-Sayed M.","family":"El-Alfy","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,13]]},"reference":[{"key":"11271_CR1","doi-asserted-by":"crossref","unstructured":"Ab Aziz NA, K T, Ismail SNMS, Hasnul MA, Ab Aziz K, Ibrahim SZ, Abd Aziz A, Raja JE (2023) Asian affective and emotional state (a2es) dataset of ecg and ppg for affective computing research. Algorithms 16(3):130","DOI":"10.3390\/a16030130"},{"issue":"3","key":"11271_CR2","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1109\/TAFFC.2015.2392932","volume":"6","author":"MK Abadi","year":"2015","unstructured":"Abadi MK, Subramanian R, Kia SM, Avesani P, Patras I, Sebe N (2015) Decaf: meg-based multimodal database for decoding affective physiological responses. IEEE Trans Affect Comput 6(3):209\u2013222","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR3","unstructured":"Abburi H, Akkireddy ESA, Gangashetty SV, Mamidi R (2016) Multimodal sentiment analysis of telugu songs. In: Proceedings of the 25th international joint conference on artificial intelligence, p\u00a048"},{"key":"11271_CR4","first-page":"407","volume":"401","author":"S Abrilian","year":"2005","unstructured":"Abrilian S, Devillers L, Buisine S, Martin JC (2005) Emotv1: annotation of real-life emotions for the specification of multimodal affective interfaces. HCI Int 401:407\u2013408","journal-title":"HCI Int"},{"issue":"7","key":"11271_CR5","doi-asserted-by":"crossref","first-page":"e12189","DOI":"10.1002\/eng2.12189","volume":"2","author":"FA Acheampong","year":"2020","unstructured":"Acheampong FA, Wenyu C, Nunoo-Mensah H (2020) Text-based emotion detection: advances, challenges, and opportunities. Eng Rep 2(7):e12189","journal-title":"Eng Rep"},{"issue":"9","key":"11271_CR6","doi-asserted-by":"crossref","first-page":"1773","DOI":"10.1038\/s41591-022-01981-2","volume":"28","author":"JN Acosta","year":"2022","unstructured":"Acosta JN, Falcone GJ, Rajpurkar P, Topol EJ (2022) Multimodal Biomedical Ai. Nat Med 28(9):1773\u20131784","journal-title":"Nat Med"},{"key":"11271_CR7","unstructured":"Adewumi T, Alkhaled L, Gurung N, van Boven G, Pagliai I (2024) Fairness and bias in multimodal ai: A survey. arXiv preprint arXiv:2406.19097"},{"key":"11271_CR8","doi-asserted-by":"crossref","unstructured":"Afzal S, Robinson P (2011) Natural affect data: collection and annotation. In: New perspectives on affect and learning technologies, Springer, pp 55\u201370","DOI":"10.1007\/978-1-4419-9625-1_5"},{"key":"11271_CR9","first-page":"200171","volume":"17","author":"N Ahmed","year":"2023","unstructured":"Ahmed N, Al Aghbari Z, Girija S (2023) A systematic survey on multimodal emotion recognition using learning algorithms. Intell Syst with Appl 17:200171","journal-title":"Intell Syst with Appl"},{"key":"11271_CR10","doi-asserted-by":"crossref","unstructured":"Ahuja K (2024) Emotion ai in healthcare: application, challenges, and future directions. In: Emotional AI and Human-AI interactions in social networking, Elsevier, pp 131\u2013146","DOI":"10.1016\/B978-0-443-19096-4.00011-0"},{"key":"11271_CR11","doi-asserted-by":"crossref","first-page":"56","DOI":"10.1016\/j.specom.2019.12.001","volume":"116","author":"MB Ak\u00e7ay","year":"2020","unstructured":"Ak\u00e7ay MB, O\u011fuz K (2020) Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers. Speech Commun 116:56\u201376","journal-title":"Speech Commun"},{"key":"11271_CR12","doi-asserted-by":"crossref","unstructured":"Al-Azani S, El-Alfy ESM (2018) Combining emojis with Arabic textual features for sentiment classification. In: Proceedings of the 9th IEEE international conference on information and communication systems (ICICS), pp 139\u2013144","DOI":"10.1109\/IACS.2018.8355456"},{"key":"11271_CR13","doi-asserted-by":"crossref","unstructured":"Al-Azani S, El-Alfy ESM (2019a) Audio-textual Arabic dialect identification for opinion mining videos. In: IEEE symposium series on computational intelligence (SSCI), pp 2470\u20132475","DOI":"10.1109\/SSCI44817.2019.9003031"},{"key":"11271_CR14","doi-asserted-by":"crossref","unstructured":"Al-Azani S, El-Alfy ESM (2019b) Multimodal sentiment and gender classification for video logs. In: Proceedings of the 11th international conference on agents and artificial intelligence (ICCART)","DOI":"10.5220\/0007711409070914"},{"key":"11271_CR15","doi-asserted-by":"crossref","first-page":"136843","DOI":"10.1109\/ACCESS.2020.3011977","volume":"8","author":"S Al-Azani","year":"2020","unstructured":"Al-Azani S, El-Alfy ESM (2020) Enhanced video analytics for sentiment analysis based on fusing textual, auditory and visual information. IEEE Access 8:136843\u2013136857","journal-title":"IEEE Access"},{"key":"11271_CR16","doi-asserted-by":"crossref","first-page":"121031","DOI":"10.1109\/ACCESS.2021.3108502","volume":"9","author":"S Al-Azani","year":"2021","unstructured":"Al-Azani S, El-Alfy ESM (2021) Early and late fusion of emojis and text to enhance opinion mining. IEEE Access 9:121031\u2013121045","journal-title":"IEEE Access"},{"key":"11271_CR17","doi-asserted-by":"crossref","first-page":"12262","DOI":"10.1109\/ACCESS.2022.3146008","volume":"10","author":"S Al-Azani","year":"2022","unstructured":"Al-Azani S, Sait SM, Al-Utaibi KA (2022) A comprehensive literature review on children\u2019s databases for machine learning applications. IEEE Access 10:12262\u201312285","journal-title":"IEEE Access"},{"key":"11271_CR18","doi-asserted-by":"crossref","first-page":"101951","DOI":"10.1016\/j.bspc.2020.101951","volume":"60","author":"TB Alakus","year":"2020","unstructured":"Alakus TB, Gonen M, Turkoglu I (2020) Database for an emotion recognition system based on eeg signals and various computer games-gameemo. Biomed Sig Process Control 60:101951","journal-title":"Biomed Sig Process Control"},{"key":"11271_CR19","doi-asserted-by":"crossref","unstructured":"Alm CO, Roth D, Sproat R (2005) Emotions from text: machine learning for text-based emotion prediction. In: Proceedings of the conference on human language technology and empirical methods in natural language processing, Association for computational linguistics, pp 579\u2013586","DOI":"10.3115\/1220575.1220648"},{"key":"11271_CR20","unstructured":"Alowisheq A, Alrajeh A, Alrowithi S, Tamran AB, Ibrahim A, Aloraini R, Alnajim R, Alkahtani R, Almuasaad R, Alrasheed S, Alsubaie S, Alonaizan Y (2023) Sada - sba & sdaia audio dataset for arabic. https:\/\/www.kaggle.com\/datasets\/sdaiancai\/sada2022, accessed: 2023-11-20"},{"key":"11271_CR21","doi-asserted-by":"crossref","unstructured":"Alqarafi AS, Adeel A, Gogate M, Dashitpour K, Hussain A, Durrani T (2017) Toward\u2019s arabic multi-modal sentiment analysis. In: Proceedings of the international conference in communications, signal processing, and systems, Springer, pp 2378\u20132386","DOI":"10.1007\/978-981-10-6571-2_290"},{"key":"11271_CR22","doi-asserted-by":"crossref","unstructured":"Alvarez-Gonzalez N, Kaltenbrunner A, G\u00f3mez V (2021) Uncovering the limits of text-based emotion detection. arXiv preprint arXiv:2109.01900","DOI":"10.18653\/v1\/2021.findings-emnlp.219"},{"issue":"1","key":"11271_CR23","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1007\/s10462-023-10631-z","volume":"57","author":"R Archana","year":"2024","unstructured":"Archana R, Jeevaraj PE (2024) Deep learning models for digital image processing: a review. Artif Intell Rev 57(1):11","journal-title":"Artif Intell Rev"},{"key":"11271_CR24","doi-asserted-by":"crossref","unstructured":"Artstein R (2017) Inter-annotator agreement. Handbook of linguistic annotation pp 297\u2013313","DOI":"10.1007\/978-94-024-0881-2_11"},{"key":"11271_CR25","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.specom.2022.03.002","volume":"140","author":"BT Atmaja","year":"2022","unstructured":"Atmaja BT, Sasou A, Akagi M (2022) Survey on bimodal speech emotion recognition from acoustic and linguistic information fusion. Speech Commun 140:11\u201328","journal-title":"Speech Commun"},{"key":"11271_CR26","doi-asserted-by":"crossref","unstructured":"Azani SA, El-Alfy ESM (2019) Multimodal age-group recognition for opinion video logs using ensemble of neural networks. Int J of Adv Comput Sci and Appl 10(4)","DOI":"10.14569\/IJACSA.2019.0100445"},{"key":"11271_CR27","doi-asserted-by":"crossref","unstructured":"Ba S, Hu X (2023) Measuring emotions in education using wearable devices: A systematic review. Computers & Education p 104797","DOI":"10.1016\/j.compedu.2023.104797"},{"key":"11271_CR28","doi-asserted-by":"crossref","unstructured":"Badawi S, Kazemi A, Rezaie V (2024) Kurdisent: a corpus for kurdish sentiment analysis. Language Resources and Evaluation pp 1\u201320","DOI":"10.1007\/s10579-023-09716-6"},{"key":"11271_CR29","doi-asserted-by":"crossref","first-page":"405","DOI":"10.1007\/s12652-020-01985-1","volume":"12","author":"A Bagirathan","year":"2021","unstructured":"Bagirathan A, Selvaraj J, Gurusamy A, Das H (2021) Recognition of positive and negative valence states in children with autism spectrum disorder (asd) using discrete wavelet transform (dwt) analysis of electrocardiogram signals (ecg). J Ambient Intell Humaniz Comput 12:405\u2013416","journal-title":"J Ambient Intell Humaniz Comput"},{"key":"11271_CR30","doi-asserted-by":"crossref","unstructured":"Baltru\u0161aitis T, Ahuja C, Morency LP (2018) Challenges and applications in multimodal machine learning. The Handbook of Multimodal-Multisensor Interfaces: Signal Processing, Architectures, and Detection of Emotion and Cognition-Volume 2:17\u201348","DOI":"10.1145\/3107990.3107993"},{"issue":"2","key":"11271_CR31","doi-asserted-by":"crossref","first-page":"423","DOI":"10.1109\/TPAMI.2018.2798607","volume":"41","author":"T Baltru\u0161aitis","year":"2018","unstructured":"Baltru\u0161aitis T, Ahuja C, Morency LP (2018) Multimodal machine learning: a survey and taxonomy. IEEE Trans Pattern Anal Mach Intell 41(2):423\u2013443","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11271_CR32","first-page":"271","volume":"2010","author":"T B\u00e4nziger","year":"2010","unstructured":"B\u00e4nziger T, Scherer KR (2010) Introducing the geneva multimodal emotion portrayal (gemep) corpus. Blueprint for affective computing: A sourcebook 2010:271\u201394","journal-title":"Blueprint for affective computing: a sourcebook"},{"issue":"6","key":"11271_CR33","first-page":"47","volume":"8","author":"L Bastida","year":"2024","unstructured":"Bastida L, Sillaurren S, Loizaga E, Tom\u00e9 E, Moya A (2024) Exploring human emotions: a virtual reality-based experimental approach integrating physiological and facial analysis. Multimodal Tech Int 8(6):47","journal-title":"Multimodal Tech Int"},{"issue":"1","key":"11271_CR34","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1109\/TAFFC.2015.2396531","volume":"6","author":"Y Baveye","year":"2015","unstructured":"Baveye Y, Dellandrea E, Chamaret C, Chen L (2015) Liris-accede: a video database for affective content analysis. IEEE Trans Affect Comput 6(1):43\u201355","journal-title":"IEEE Trans Affect Comput"},{"issue":"3","key":"11271_CR35","doi-asserted-by":"crossref","first-page":"2325","DOI":"10.1007\/s10462-022-10215-3","volume":"56","author":"R Bensoltane","year":"2023","unstructured":"Bensoltane R, Zaki T (2023) Aspect-based sentiment analysis: an overview in the use of arabic language. Artif Intell Rev 56(3):2325\u20132363","journal-title":"Artif Intell Rev"},{"issue":"22","key":"11271_CR36","doi-asserted-by":"crossref","first-page":"8534","DOI":"10.3390\/ijerph17228534","volume":"17","author":"M Blanco-Ruiz","year":"2020","unstructured":"Blanco-Ruiz M, Sainz-de Baranda C, Guti\u00e9rrez-Mart\u00edn L, Romero-Perales E, L\u00f3pez-Ongil C (2020) Emotion elicitation under audiovisual stimuli reception: should artificial intelligence consider the gender perspective? Int J Environ Res Public Health 17(22):8534","journal-title":"Int J Environ Res Public Health"},{"key":"11271_CR37","doi-asserted-by":"crossref","unstructured":"Boccignone G, Conte D, Cuculo V, Lanzarotti R (2017) Amhuse: a multimodal dataset for humour sensing. In: Proceedings of the 19th ACM international conference on multimodal interaction, pp 438\u2013445","DOI":"10.1145\/3136755.3136806"},{"key":"11271_CR38","doi-asserted-by":"crossref","unstructured":"Bohus D, Rudnicky AI (2008) Sorry, i didn\u2019t catch that! In: Recent trends in discourse and dialogue, Springer, pp 123\u2013154","DOI":"10.1007\/978-1-4020-6821-8_6"},{"key":"11271_CR39","doi-asserted-by":"crossref","unstructured":"Borth D, Ji R, Chen T, Breuel T, Chang SF (2013) Large-scale visual sentiment ontology and detectors using adjective noun pairs. In: Proceedings of the 21st ACM international conference on multimedia, pp 223\u2013232","DOI":"10.1145\/2502081.2502282"},{"key":"11271_CR40","doi-asserted-by":"crossref","first-page":"140990","DOI":"10.1109\/ACCESS.2019.2944001","volume":"7","author":"PJ Bota","year":"2019","unstructured":"Bota PJ, Wang C, Fred AL, Da Silva HP (2019) A review, current challenges, and future possibilities on emotion recognition using machine learning and physiological signals. IEEE Access 7:140990\u2013141020","journal-title":"IEEE Access"},{"key":"11271_CR41","doi-asserted-by":"crossref","unstructured":"Bousmalis K, Morency LP, Pantic M (2011) Modeling hidden dynamics of multimodal cues for spontaneous agreement and disagreement recognition. In: Face and gesture 2011, IEEE, pp 746\u2013752","DOI":"10.1109\/FG.2011.5771341"},{"issue":"1","key":"11271_CR42","doi-asserted-by":"crossref","first-page":"49","DOI":"10.1016\/0005-7916(94)90063-9","volume":"25","author":"MM Bradley","year":"1994","unstructured":"Bradley MM, Lang PJ (1994) Measuring emotion: the self-assessment manikin and the semantic differential. J Behav Ther Exp Psychiatry 25(1):49\u201359","journal-title":"J Behav Ther Exp Psychiatry"},{"issue":"6","key":"11271_CR43","doi-asserted-by":"crossref","first-page":"641","DOI":"10.1016\/j.ijhcs.2013.02.003","volume":"71","author":"J Broekens","year":"2013","unstructured":"Broekens J, Brinkman WP (2013) Affectbutton: a method for reliable and valid affective self-report. Int J Hum Comput Stud 71(6):641\u2013667","journal-title":"Int J Hum Comput Stud"},{"key":"11271_CR44","unstructured":"Brugman H, Russel A, Nijmegen X (2004) Annotating multi-media\/multi-modal resources with elan. In: LREC, Citeseer"},{"key":"11271_CR45","first-page":"2925","volume":"2024","author":"P Bujnowski","year":"2024","unstructured":"Bujnowski P, Kuzma B, Paziewski B, Rutkowski J, Marhula J, Bordzicka Z, Andruszkiewicz P (2024) Samsemo: new dataset for multilingual and multimodal emotion recognition. Proc. Interspeech 2024:2925\u20132929","journal-title":"Proc. Interspeech"},{"key":"11271_CR46","doi-asserted-by":"publisher","unstructured":"Burkhardt F, Paeschke A, Rolfes M, Sendlmeier WF, Weiss B (2005) A database of German emotional speech. In: Proc. interspeech 2005, pp 1517\u20131520, https:\/\/doi.org\/10.21437\/Interspeech.2005-446","DOI":"10.21437\/Interspeech.2005-446"},{"issue":"4","key":"11271_CR47","doi-asserted-by":"crossref","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","volume":"42","author":"C Busso","year":"2008","unstructured":"Busso C, Bulut M, Lee CC, Kazemzadeh A, Mower E, Kim S, Chang JN, Lee S, Narayanan SS (2008) IEMOCAP: interactive emotional dyadic motion capture database. Lang Resour Eval 42(4):335\u2013359","journal-title":"Lang Resour Eval"},{"issue":"1","key":"11271_CR48","doi-asserted-by":"crossref","first-page":"67","DOI":"10.1109\/TAFFC.2016.2515617","volume":"8","author":"C Busso","year":"2016","unstructured":"Busso C, Parthasarathy S, Burmania A, AbdelWahab M, Sadoughi N, Provost EM (2016) Msp-improv: an acted corpus of dyadic interactions to study emotion perception. IEEE Trans Affect Comput 8(1):67\u201380","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR49","doi-asserted-by":"crossref","first-page":"103058","DOI":"10.1016\/j.cose.2022.103058","volume":"125","author":"MA Butt","year":"2023","unstructured":"Butt MA, Qayyum A, Ali H, Al-Fuqaha A, Qadir J (2023) Towards secure private and trustworthy human-centric embedded machine learning: an emotion-aware facial recognition case study. Comp Security 125:103058","journal-title":"Comp Security"},{"issue":"5","key":"11271_CR50","doi-asserted-by":"crossref","first-page":"2455","DOI":"10.3390\/s23052455","volume":"23","author":"Y Cai","year":"2023","unstructured":"Cai Y, Li X, Li J (2023) Emotion recognition using different sensors, emotion models, methods and datasets: a comprehensive review. Sensors 23(5):2455","journal-title":"Sensors"},{"issue":"2","key":"11271_CR51","doi-asserted-by":"crossref","first-page":"102","DOI":"10.1109\/MIS.2016.31","volume":"31","author":"E Cambria","year":"2016","unstructured":"Cambria E (2016) Affective computing and sentiment analysis. IEEE Intell Syst 31(2):102\u2013107","journal-title":"IEEE Intell Syst"},{"issue":"11","key":"11271_CR52","doi-asserted-by":"crossref","first-page":"5291","DOI":"10.3390\/s23115291","volume":"23","author":"S Cano","year":"2023","unstructured":"Cano S, D\u00edaz-Arancibia J, Arango-L\u00f3pez J, Libreros JE, Garc\u00eda M (2023) Design path for a social robot for emotional communication for children with autism spectrum disorder (asd). Sensors 23(11):5291","journal-title":"Sensors"},{"key":"11271_CR53","doi-asserted-by":"crossref","unstructured":"Caridakis G, Castellano G, Kessous L, Raouzaiou A, Malatesta L, Asteriadis S, Karpouzis K (2007) Multimodal emotion recognition from expressive faces, body gestures and speech. In: IFIP international conference on artificial intelligence applications and innovations, Springer, pp 375\u2013388","DOI":"10.1007\/978-0-387-74161-1_41"},{"key":"11271_CR54","unstructured":"Caridakis G, Wagner J, Raouzaiou A, Curto Z, Andr\u00e9 E, Karpouzis K (2010) A multimodal corpus for gesture expressivity analysis. Multimodal Corpora: Advances in Capturing, Coding and Analyzing Multimodality p\u00a080"},{"key":"11271_CR55","unstructured":"Castellano B (2023) PySceneDetect. https:\/\/github.com\/Breakthrough\/PySceneDetect, accessed: January, 10, 2024"},{"key":"11271_CR56","doi-asserted-by":"crossref","unstructured":"Castro S, Hazarika D, P\u00e9rez-Rosas V, Zimmermann R, Mihalcea R, Poria S (2019) Towards multimodal sarcasm detection (an _obviously_ perfect paper). In: Proceedings of the 57th annual meeting of the association for computational linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Florence, Italy","DOI":"10.18653\/v1\/P19-1455"},{"issue":"24","key":"11271_CR57","doi-asserted-by":"crossref","first-page":"5462","DOI":"10.3390\/app9245462","volume":"9","author":"P Chakriswaran","year":"2019","unstructured":"Chakriswaran P, Vincent DR, Srinivasan K, Sharma V, Chang CY, Reina DG (2019) Emotion ai-driven sentiment analysis: a survey, future research directions, and open issues. Appl Sci 9(24):5462","journal-title":"Appl Sci"},{"issue":"1","key":"11271_CR58","doi-asserted-by":"crossref","first-page":"749","DOI":"10.1007\/s10462-022-10183-8","volume":"56","author":"JYL Chan","year":"2023","unstructured":"Chan JYL, Bea KT, Leow SMH, Phoong SW, Cheng WK (2023) State of the art: a review of sentiment analysis based on sequential transfer learning. Artif Intell Rev 56(1):749\u2013780","journal-title":"Artif Intell Rev"},{"key":"11271_CR59","doi-asserted-by":"crossref","unstructured":"Chandrasekaran G, Nguyen TN, Hemanth DJ (2021) Multimodal sentimental analysis for social media applications: A comprehensive review. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 11(5):e1415","DOI":"10.1002\/widm.1415"},{"key":"11271_CR60","unstructured":"Chaptoukaev H, Strizhkova V, Panariello M, D\u2019alpaos B, Reka A, Manera V, Th\u00fcmmler S, Ismailova E, Evans N, Bremond FF, et al (2023) Stressid: a multimodal dataset for stress identification. In: Proceedings of 37th conference on neural information processing systems (NeurIPS)"},{"issue":"2","key":"11271_CR61","doi-asserted-by":"crossref","first-page":"103193","DOI":"10.1016\/j.ipm.2022.103193","volume":"60","author":"D Chen","year":"2023","unstructured":"Chen D, Su W, Wu P, Hua B (2023) Joint multimodal sentiment analysis based on information relevance. Inf Process Manag 60(2):103193","journal-title":"Inf Process Manag"},{"key":"11271_CR62","doi-asserted-by":"crossref","unstructured":"Chen J, Wang C, Wang K, Yin C, Zhao C, Xu T, Zhang X, Huang Z, Liu M, Yang T (2021) Heu emotion: a large-scale database for multimodal emotion recognition in the wild. Neural Comput and Appl pp 1\u201317","DOI":"10.1007\/s00521-020-05616-w"},{"issue":"1","key":"11271_CR63","doi-asserted-by":"crossref","first-page":"14","DOI":"10.1109\/MSMC.2019.2929312","volume":"6","author":"M Chen","year":"2020","unstructured":"Chen M, Jiang Y, Cao Y, Zomaya AY (2020) Creativebioman: a brain-and body-wearable, computing-based, creative gaming system. IEEE Syst, Man, and Cybern Magaz 6(1):14\u201322","journal-title":"IEEE Syst, Man, and Cybern Magaz"},{"key":"11271_CR64","unstructured":"Chen SY, Hsu CC, Kuo CC, Ku LW, et al (2018) Emotionlines: An emotion corpus of multi-party conversations. arXiv preprint arXiv:1802.08379"},{"key":"11271_CR65","doi-asserted-by":"crossref","first-page":"e1901","DOI":"10.7717\/peerj-cs.1901","volume":"10","author":"P Cherukuru","year":"2024","unstructured":"Cherukuru P, Mustafa MB (2024) Cnn-based noise reduction for multi-channel speech enhancement system with discrete wavelet transform (dwt) preprocessing. PeerJ Comp Sci 10:e1901","journal-title":"PeerJ Comp Sci"},{"key":"11271_CR66","doi-asserted-by":"crossref","unstructured":"Chou HC, Lin WC, Chang LC, Li CC, Ma HP, Lee CC (2017) Nnime: The nthu-ntua chinese interactive multimodal emotion corpus. In: Seventh international conference on affective computing and intelligent interaction (ACII), pp 292\u2013298","DOI":"10.1109\/ACII.2017.8273615"},{"issue":"1","key":"11271_CR67","first-page":"4","volume":"46","author":"JT Colorado","year":"2012","unstructured":"Colorado JT, Eberle J (2012) Student demographics and success in online learning environments. Emporia State Res Studies 46(1):4\u201310","journal-title":"Emporia State Res Studies"},{"issue":"9","key":"11271_CR68","doi-asserted-by":"crossref","first-page":"5697","DOI":"10.3390\/app13095697","volume":"13","author":"W Costa","year":"2023","unstructured":"Costa W, Talavera E, Oliveira R, Figueiredo L, Teixeira JM, Lima JP, Teichrieb V (2023) A survey on datasets for emotion recognition from vision: limitations and in-the-wild applicability. Appl Sci 13(9):5697","journal-title":"Appl Sci"},{"issue":"2","key":"11271_CR69","doi-asserted-by":"crossref","first-page":"124","DOI":"10.1016\/j.tics.2020.11.004","volume":"25","author":"AS Cowen","year":"2021","unstructured":"Cowen AS, Keltner D (2021) Semantic space theory: a computational approach to emotion. Trends Cogn Sci 25(2):124\u2013136","journal-title":"Trends Cogn Sci"},{"key":"11271_CR70","unstructured":"Cristea D, Pistol I, Boghiu S, Bibiri AD, G\u00eefu D, Scutelnicu A, Plamada-Onofrei M, Trandabat D, Bugeag G (2020) Cobiliro: A research platform for bimodal corpora. In: Proceedings of the 1st international workshop on language technology platforms, pp 22\u201327"},{"issue":"2","key":"11271_CR71","doi-asserted-by":"crossref","first-page":"254","DOI":"10.3390\/math13020254","volume":"13","author":"JA Cruz-Vazquez","year":"2025","unstructured":"Cruz-Vazquez JA, Montiel-P\u00e9rez JY, Romero-Herrera R, Rubio-Espino E (2025) Emotion recognition from eeg signals using advanced transformations and deep learning. Mathematics 13(2):254","journal-title":"Mathematics"},{"key":"11271_CR72","doi-asserted-by":"crossref","unstructured":"Davis GM (2018) Noise reduction in speech applications. CRC Press","DOI":"10.1201\/9781315220109"},{"key":"11271_CR73","doi-asserted-by":"crossref","unstructured":"Degottex G, Kane J, Drugman T, Raitio T, Scherer S (2014) Covarep-a collaborative voice analysis repository for speech technologies. In: IEEE international conference on acoustics, speech and signal processing (icassp), pp 960\u2013964","DOI":"10.1109\/ICASSP.2014.6853739"},{"key":"11271_CR74","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1007\/s10994-012-5285-8","volume":"88","author":"K Dembczy\u0144ski","year":"2012","unstructured":"Dembczy\u0144ski K, Waegeman W, Cheng W, H\u00fcllermeier E (2012) On label dependence and loss minimization in multi-label classification. Mach Learn 88:5\u201345","journal-title":"Mach Learn"},{"issue":"1","key":"11271_CR75","doi-asserted-by":"crossref","first-page":"17","DOI":"10.1016\/j.artmed.2015.03.006","volume":"64","author":"K Denecke","year":"2015","unstructured":"Denecke K, Deng Y (2015) Sentiment analysis in medical settings: new opportunities and challenges. Artif Intell Med 64(1):17\u201327","journal-title":"Artif Intell Med"},{"issue":"4","key":"11271_CR76","doi-asserted-by":"crossref","first-page":"407","DOI":"10.1016\/j.neunet.2005.03.007","volume":"18","author":"L Devillers","year":"2005","unstructured":"Devillers L, Vidrascu L, Lamel L (2005) Challenges in real-life emotion annotation and machine learning based detection. Neural Netw 18(4):407\u2013422","journal-title":"Neural Netw"},{"key":"11271_CR77","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1109\/MMUL.2012.26","volume":"3","author":"A Dhall","year":"2012","unstructured":"Dhall A, Goecke R, Lucey S, Gedeon T (2012) Collecting large, richly annotated facial-expression databases from movies. IEEE Multimedia 19(3):34\u201341","journal-title":"IEEE Multimedia"},{"key":"11271_CR79","doi-asserted-by":"crossref","unstructured":"Dhiman B, Kantaria D, Khan K, Sankaran HK (2023) Artificial intelligence-based learning toys: Exploring the role of tangram as a tool to develop spatial learning among children. In: International conference on research into design, Springer, pp 1023\u20131031","DOI":"10.1007\/978-981-99-0293-4_82"},{"issue":"1","key":"11271_CR80","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1177\/17540739211072803","volume":"14","author":"K Diconne","year":"2021","unstructured":"Diconne K, Kountouriotis G, Paltoglou AE, Parker A, Hostler TJ (2021) Kapodi-the searchable database of 364 available emotional stimuli sets. Emotion Rev 14(1):84\u201395","journal-title":"Emotion Rev"},{"issue":"3","key":"11271_CR81","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/2682899","volume":"47","author":"SK D\u2019mello","year":"2015","unstructured":"D\u2019mello SK, Kory J (2015) A review and meta-analysis of multimodal affect detection systems. ACM Comp Surv (CSUR) 47(3):1\u201336","journal-title":"ACM Comp Surv (CSUR)"},{"key":"11271_CR82","unstructured":"Domnich A, Anbarjafari G (2021) Responsible ai: Gender bias assessment in emotion recognition. arXiv preprint arXiv:2103.11436"},{"key":"11271_CR83","unstructured":"Douglas-Cowie E (1978) Linguistic code-switching in a northern irish village. In: Sociolinguistic patterns in British English, Edward Arnold, pp 37\u201351"},{"key":"11271_CR84","unstructured":"Douglas-Cowie E, Cowie R, Schr\u00f6der M (2000) A new emotion database: considerations, sources and scope. In: ISCA tutorial and research workshop (ITRW) on speech and emotion"},{"issue":"1\u20132","key":"11271_CR85","doi-asserted-by":"crossref","first-page":"33","DOI":"10.1016\/S0167-6393(02)00070-5","volume":"40","author":"E Douglas-Cowie","year":"2003","unstructured":"Douglas-Cowie E, Campbell N, Cowie R, Roach P (2003) Emotional speech: towards a new generation of databases. Speech Commun 40(1\u20132):33\u201360","journal-title":"Speech Commun"},{"key":"11271_CR86","doi-asserted-by":"crossref","unstructured":"Douglas-Cowie E, Cowie R, Sneddon I, Cox C, Lowry O, Mcrorie M, Martin JC, Devillers L, Abrilian S, Batliner A, et al (2007) The humaine database: Addressing the collection and annotation of naturalistic and induced emotional data. In: International conference on affective computing and intelligent interaction, Springer, pp 488\u2013500","DOI":"10.1007\/978-3-540-74889-2_43"},{"key":"11271_CR87","doi-asserted-by":"crossref","unstructured":"Doyran M, Schimmel A, Baki P, Ergin K, T\u00fcrkmen B, Salah AA, Bakkes SC, Kaya H, Poppe R, Salah AA (2021) Mumbai: multi-person, multimodal board game affect and interaction analysis dataset. J Multimodal User Interfaces 15: 373-391","DOI":"10.1007\/s12193-021-00364-0"},{"issue":"3\u20134","key":"11271_CR88","doi-asserted-by":"crossref","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman P (1992) An argument for basic emotions. Cognition Emotion 6(3\u20134):169\u2013200","journal-title":"Cognition Emotion"},{"key":"11271_CR89","unstructured":"Ekman P (2002) Facial action coding system (facs). A human face"},{"key":"11271_CR90","first-page":"599","volume":"31","author":"P Ekman","year":"1984","unstructured":"Ekman P, Friesen W (1984) Unmasking the face. palo alto. CA: Consul Psychol Press 31:599\u2013612","journal-title":"CA: Consul Psychol Press"},{"key":"11271_CR91","doi-asserted-by":"crossref","unstructured":"Ekman P, Friesen WV (1978) Manual for the facial action coding system. Consulting Psychologists Press","DOI":"10.1037\/t27734-000"},{"key":"11271_CR92","doi-asserted-by":"crossref","unstructured":"Ellis JG, Jou B, Chang SF (2014) Why we watch the news: a dataset for exploring sentiment in broadcast video news. In: Proceedings of the 16th international conference on multimodal interaction, ACM, pp 104\u2013111","DOI":"10.1145\/2663204.2663237"},{"key":"11271_CR93","doi-asserted-by":"crossref","unstructured":"Eyben F, W\u00f6llmer M, Schuller B (2009) OpenEAR-introducing the munich open-source emotion and affect recognition toolkit. In: Proceedings of the 3rd IEEE international conference on affective computing and intelligent interaction and workshops, pp 1\u20136","DOI":"10.1109\/ACII.2009.5349350"},{"key":"11271_CR94","doi-asserted-by":"crossref","unstructured":"Fanni SC, Febi M, Aghakhanyan G, Neri E (2023) Natural language processing. In: Introduction to artificial intelligence, Springer, pp 87\u201399","DOI":"10.1007\/978-3-031-25928-9_5"},{"key":"11271_CR95","doi-asserted-by":"crossref","first-page":"1234162","DOI":"10.3389\/fnins.2023.1234162","volume":"17","author":"B Fu","year":"2023","unstructured":"Fu B, Gu C, Fu M, Xia Y, Liu Y (2023) A novel feature fusion network for multimodal emotion recognition from eeg and eye movement signals. Front Neurosci 17:1234162","journal-title":"Front Neurosci"},{"key":"11271_CR96","unstructured":"Furr RM (2021) Psychometrics: an introduction. SAGE Publications"},{"issue":"5","key":"11271_CR97","doi-asserted-by":"crossref","first-page":"829","DOI":"10.1162\/neco_a_01273","volume":"32","author":"J Gao","year":"2020","unstructured":"Gao J, Li P, Chen Z, Zhang J (2020) A survey on deep learning for multimodal data fusion. Neural Comput 32(5):829\u2013864","journal-title":"Neural Comput"},{"key":"11271_CR98","doi-asserted-by":"crossref","first-page":"105519","DOI":"10.1016\/j.imavis.2025.105519","volume":"158","author":"M Gao","year":"2025","unstructured":"Gao M, Sun J, Li Q, Khan MA, Shang J, Zhu X, Jeon G (2025) Towards trustworthy image super-resolution via symmetrical and recursive artificial neural network. Image Vis Comput 158:105519","journal-title":"Image Vis Comput"},{"key":"11271_CR99","doi-asserted-by":"crossref","first-page":"107660","DOI":"10.1016\/j.dib.2021.107660","volume":"39","author":"Z Gao","year":"2021","unstructured":"Gao Z, Cui X, Wan W, Zheng W, Gu Z (2021) Ecsmp: a dataset on emotion, cognition, sleep, and multi-model physiological signals. Data Brief 39:107660","journal-title":"Data Brief"},{"key":"11271_CR101","unstructured":"Gnjatovi\u0107 M, R\u00f6sner D (2008) The nimitek corpus of affected behavior in human-machine interaction. In: Programme of the Workshop on Corpora for Research on Emotion and Affect 10:5"},{"key":"11271_CR102","unstructured":"Go A, Bhayani R, Huang L (2009) Twitter sentiment classification using distant supervision. CS224N project report, Stanford Technical Report"},{"key":"11271_CR103","doi-asserted-by":"crossref","unstructured":"Goodfellow IJ, Erhan D, Carrier PL, Courville A, Mirza M, Hamner B, Cukierski W, Tang Y, Thaler D, Lee DH, et\u00a0al (2013) Challenges in representation learning: A report on three machine learning contests. In: Neural information processing: 20th international conference, ICONIP 2013, Daegu, Korea, November 3-7, 2013. Proceedings, Part III 20, Springer, pp 117\u2013124","DOI":"10.1007\/978-3-642-42051-1_16"},{"issue":"3","key":"11271_CR104","first-page":"494","volume":"38","author":"S Govindaraj","year":"2016","unstructured":"Govindaraj S, Gopalakrishnan K (2016) Intensified sentiment analysis of customer product reviews using acoustic and textual features. ETRI J 38(3):494\u2013501","journal-title":"ETRI J"},{"key":"11271_CR105","doi-asserted-by":"crossref","unstructured":"Grimm M, Kroschel K, Narayanan S (2008) The vera am mittag german audio-visual emotional speech database. In: IEEE international conference on multimedia and expo, pp 865\u2013868","DOI":"10.1109\/ICME.2008.4607572"},{"key":"11271_CR106","doi-asserted-by":"crossref","unstructured":"Gunes H, Piccardi M (2006) A bimodal face and body gesture database for automatic analysis of human nonverbal affective behavior. In: 18th IEEE International conference on pattern recognition (ICPR\u201906)","DOI":"10.1109\/ICPR.2006.39"},{"key":"11271_CR107","doi-asserted-by":"crossref","unstructured":"Guo S, et\u00a0al (2024) Emotional expression in artworks and psychological reactions of audiences. Journal of Art, Culture and Philosophical Studies 1(3)","DOI":"10.70767\/jacps.v1i3.388"},{"key":"11271_CR108","doi-asserted-by":"crossref","unstructured":"Hasnul MA, Ab\u00a0Aziz NA, Abd\u00a0Aziz A (2021a) Evaluation of teap and aubt as ecg\u2019s feature extraction toolbox for emotion recognition system. In: IEEE 9th conference on systems, process and control (ICSPC 2021), pp 52\u201357","DOI":"10.1109\/ICSPC53359.2021.9689133"},{"issue":"15","key":"11271_CR109","doi-asserted-by":"crossref","first-page":"5015","DOI":"10.3390\/s21155015","volume":"21","author":"MA Hasnul","year":"2021","unstructured":"Hasnul MA, Aziz NAA, Alelyani S, Mohana M, Aziz AA (2021) Electrocardiogram-based emotion recognition systems and their applications in healthcare-a review. Sensors 21(15):5015","journal-title":"Sensors"},{"key":"11271_CR110","doi-asserted-by":"crossref","first-page":"108339","DOI":"10.1016\/j.engappai.2024.108339","volume":"133","author":"S Hazmoune","year":"2024","unstructured":"Hazmoune S, Bougamouza F (2024) Using transformers for multimodal emotion recognition: Taxonomies and state of the art review. Eng Appl Artif Intell 133:108339","journal-title":"Eng Appl Artif Intell"},{"issue":"2","key":"11271_CR111","doi-asserted-by":"crossref","first-page":"156","DOI":"10.1109\/TITS.2005.848368","volume":"6","author":"JA Healey","year":"2005","unstructured":"Healey JA, Picard RW (2005) Detecting stress during real-world driving tasks using physiological sensors. IEEE Trans Intell Transp Syst 6(2):156\u2013166","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"11271_CR112","doi-asserted-by":"crossref","unstructured":"Henriksen D, Creely E, Gruber N, Leahy S (2025) Social-emotional learning and generative ai: A critical literature review and framework for teacher education. Journal of Teacher Education p 00224871251325058","DOI":"10.1177\/00224871251325058"},{"key":"11271_CR113","doi-asserted-by":"crossref","unstructured":"Hosseinzadeh M, Azhir E, Ahmed OH, Ghafour MY, Ahmed SH, Rahmani AM, Vo B (2023) Data cleansing mechanisms and approaches for big data analytics: a systematic study. Journal of Ambient Intelligence and Humanized Computing pp 1\u201313","DOI":"10.1007\/s12652-021-03590-2"},{"key":"11271_CR114","doi-asserted-by":"crossref","unstructured":"Hsu CC, Lee CM, Chou YS (2024) Drct: Saving image super-resolution away from information bottleneck. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6133\u20136142","DOI":"10.1109\/CVPRW63382.2024.00618"},{"issue":"1","key":"11271_CR115","doi-asserted-by":"crossref","first-page":"85","DOI":"10.1109\/TAFFC.2017.2781732","volume":"11","author":"YL Hsu","year":"2017","unstructured":"Hsu YL, Wang JS, Chiang WC, Hung CH (2017) Automatic ecg-based emotion recognition in music listening. IEEE Trans Affect Comput 11(1):85\u201399","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR116","doi-asserted-by":"crossref","unstructured":"Hu A, Flaxman S (2018) Multimodal sentiment analysis to explore the structure of emotions. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, pp 350\u2013358","DOI":"10.1145\/3219819.3219853"},{"key":"11271_CR117","doi-asserted-by":"crossref","unstructured":"Hu M, Liu B (2004) Mining and summarizing customer reviews. In: Proceedings of the 10th ACM SIGKDD international conference on knowledge discovery and data mining, pp 168\u2013177","DOI":"10.1145\/1014052.1014073"},{"key":"11271_CR118","doi-asserted-by":"crossref","unstructured":"Hulliyah K, Bakar NSAA, Ismail AR (2017) Emotion recognition and brain mapping for sentiment analysis: A review. In: 2017 second international conference on informatics and computing (ICIC), IEEE, pp 1\u20135","DOI":"10.1109\/IAC.2017.8280568"},{"key":"11271_CR119","doi-asserted-by":"crossref","unstructured":"Hutto CJ, Gilbert E (2014) Vader: A parsimonious rule-based model for sentiment analysis of social media text. In: Proceedings of the 8th international AAAI conference on weblogs and social media","DOI":"10.1609\/icwsm.v8i1.14550"},{"issue":"1","key":"11271_CR120","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1146\/annurev.psych.60.110707.163539","volume":"60","author":"CE Izard","year":"2009","unstructured":"Izard CE (2009) Emotion theory and research: highlights, unanswered questions, and emerging issues. Annu Rev Psychol 60(1):1\u201325","journal-title":"Annu Rev Psychol"},{"issue":"7","key":"11271_CR121","doi-asserted-by":"crossref","first-page":"2538","DOI":"10.3390\/s22072538","volume":"22","author":"P Jemio\u0142o","year":"2022","unstructured":"Jemio\u0142o P, Storman D, Mamica M, Szymkowski M, \u017babicka W, Wojtaszek-G\u0142\u00f3wka M, Lig\u0119za A (2022) Datasets for automated affect and emotion recognition from cardiovascular signals using artificial intelligence-a systematic review. Sensors 22(7):2538","journal-title":"Sensors"},{"key":"11271_CR122","doi-asserted-by":"crossref","unstructured":"Jiang Q, Chen L, Xu R, Ao X, Yang M (2019) A challenge dataset and effective models for aspect-based sentiment analysis. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 6280\u20136285","DOI":"10.18653\/v1\/D19-1654"},{"key":"11271_CR123","doi-asserted-by":"crossref","unstructured":"Jiang YG, Xu B, Xue X (2014) Predicting emotions in user-generated videos. In: Proceedings of the 28 th AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v28i1.8724"},{"key":"11271_CR124","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s10919-017-0268-x","volume":"42","author":"PN Juslin","year":"2018","unstructured":"Juslin PN, Laukka P, B\u00e4nziger T (2018) The mirror to our soul? comparisons of spontaneous and posed vocal expression of emotion. J Nonverbal Behav 42:1\u201340","journal-title":"J Nonverbal Behav"},{"key":"11271_CR125","doi-asserted-by":"crossref","first-page":"113769","DOI":"10.1109\/ACCESS.2023.3325037","volume":"11","author":"S Kakuba","year":"2023","unstructured":"Kakuba S, Poulose A, Han DS (2023) Deep learning approaches for bimodal speech emotion recognition: advancements, challenges, and a multi-learning model. IEEE Access 11:113769\u2013113789","journal-title":"IEEE Access"},{"key":"11271_CR126","doi-asserted-by":"crossref","unstructured":"Katirai A (2023) Ethical considerations in emotion recognition technologies: a review of the literature. AI and Ethics pp 1\u201322","DOI":"10.1007\/s43681-023-00307-3"},{"issue":"1","key":"11271_CR127","doi-asserted-by":"crossref","first-page":"98","DOI":"10.1109\/JBHI.2017.2688239","volume":"22","author":"S Katsigiannis","year":"2018","unstructured":"Katsigiannis S, Ramzan N (2018) Dreamer: a database for emotion recognition through eeg and ecg signals from wireless low-cost off-the-shelf devices. IEEE J Biomed Health Inform 22(1):98\u2013107","journal-title":"IEEE J Biomed Health Inform"},{"key":"11271_CR128","doi-asserted-by":"crossref","unstructured":"Kaur A, Mustafa A, Mehta L, Dhall A (2018) Prediction and localization of student engagement in the wild. In: IEEE digital image computing: techniques and applications (DICTA),pp 1\u20138","DOI":"10.1109\/DICTA.2018.8615851"},{"key":"11271_CR129","doi-asserted-by":"crossref","first-page":"108788","DOI":"10.1016\/j.asoc.2022.108788","volume":"122","author":"H Kaushik","year":"2022","unstructured":"Kaushik H, Kumar T, Bhalla K (2022) isecurehome: a deep fusion framework for surveillance of smart homes using real-time emotion recognition. Appl Soft Comput 122:108788","journal-title":"Appl Soft Comput"},{"key":"11271_CR130","doi-asserted-by":"crossref","unstructured":"Keerthika M, Abilash K, Vasanth MS, Kuralamudhu M, Abbirooban S (2024) Emotional ai: Computationally intelligent devices for education. In: Emotional AI and Human-AI interactions in social networking, Elsevier, pp 87\u201399","DOI":"10.1016\/B978-0-443-19096-4.00007-9"},{"issue":"3","key":"11271_CR131","doi-asserted-by":"crossref","first-page":"242","DOI":"10.1177\/09637214221150511","volume":"32","author":"D Keltner","year":"2023","unstructured":"Keltner D, Brooks JA, Cowen A (2023) Semantic space theory: data-driven insights into basic emotions. Curr Dir Psychol Sci 32(3):242\u2013249","journal-title":"Curr Dir Psychol Sci"},{"issue":"1","key":"11271_CR132","doi-asserted-by":"crossref","first-page":"5473","DOI":"10.1038\/s41598-025-89202-x","volume":"15","author":"M Khan","year":"2025","unstructured":"Khan M, Tran PN, Pham NT, El Saddik A, Othmani A (2025) Memocmt: multimodal emotion recognition using cross-modal transformer-based feature fusion. Sci Rep 15(1):5473","journal-title":"Sci Rep"},{"issue":"6","key":"11271_CR133","doi-asserted-by":"crossref","first-page":"9479","DOI":"10.1007\/s11042-020-10106-1","volume":"80","author":"TLB Khanh","year":"2021","unstructured":"Khanh TLB, Kim SH, Lee G, Yang HJ, Baek ET (2021) Korean video dataset for emotion recognition in the wild. Multi Tools Appl 80(6):9479\u20139492","journal-title":"Multi Tools Appl"},{"issue":"4","key":"11271_CR134","doi-asserted-by":"crossref","first-page":"9911","DOI":"10.1007\/s11042-023-16066-6","volume":"83","author":"D Khanna","year":"2024","unstructured":"Khanna D, Jindal N, Rana PS, Singh H (2024) Enhanced spatio-temporal 3d cnn for facial expression classification in videos. Multi Tools Appl 83(4):9911\u20139928","journal-title":"Multi Tools Appl"},{"issue":"1","key":"11271_CR136","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","volume":"3","author":"S Koelstra","year":"2012","unstructured":"Koelstra S, Muhl C, Soleymani M, Lee JS, Yazdani A, Ebrahimi T, Pun T, Nijholt A, Patras I (2012) Deap: a database for emotion analysis; using physiological signals. IEEE Trans Affect Comput 3(1):18\u201331","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR137","doi-asserted-by":"crossref","unstructured":"Koldijk S, Sappelli M, Verberne S, Neerincx MA, Kraaij W (2014) The swell knowledge work dataset for stress and user modeling research. In: Proceedings of the 16th international conference on multimodal interaction, pp 291\u2013298","DOI":"10.1145\/2663204.2663257"},{"issue":"1","key":"11271_CR138","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1177\/001316447003000105","volume":"30","author":"K Krippendorff","year":"1970","unstructured":"Krippendorff K (1970) Estimating the reliability, systematic error and random error of interval data. Educ Psychol Measur 30(1):61\u201370","journal-title":"Educ Psychol Measur"},{"issue":"2","key":"11271_CR139","doi-asserted-by":"crossref","first-page":"374","DOI":"10.3390\/electronics9020374","volume":"9","author":"S Kumar","year":"2020","unstructured":"Kumar S, Gahalawat M, Roy PP, Dogra DP, Kim BG (2020) Exploring impact of age and gender on sentiment analysis using machine learning. Electronics 9(2):374","journal-title":"Electronics"},{"key":"11271_CR140","doi-asserted-by":"crossref","unstructured":"Kumar S, Mondal I, Akhtar MS, Chakraborty T (2023) Explaining (sarcastic) utterances to enhance affect understanding in multimodal dialogues. Proceedings of the AAAI Conference on Artificial Intelligence 37:12986\u201312994","DOI":"10.1609\/aaai.v37i11.26526"},{"key":"11271_CR141","doi-asserted-by":"crossref","unstructured":"Kumar T, Mahrishi M, Nawaz S (2022) A review of speech sentiment analysis using machine learning. Proceedings of Trends in Electronics and Health Informatics: TEHI 2021:21\u201328","DOI":"10.1007\/978-981-16-8826-3_3"},{"issue":"3","key":"11271_CR142","doi-asserted-by":"crossref","first-page":"43","DOI":"10.3390\/bdcc5030043","volume":"5","author":"S Kusal","year":"2021","unstructured":"Kusal S, Patil S, Kotecha K, Aluvalu R, Varadarajan V (2021) Ai based emotion detection for textual big data: techniques and contribution. Big Data and Cognitive Comp 5(3):43","journal-title":"Big Data and Cognitive Comp"},{"key":"11271_CR143","unstructured":"Kutt K, Drazyk D, Jemiolo P, Bobek S, Gizycka B, Rodriguez-Fernandez V, Nalepa GJ (2019) Biraffe: Bio-reactions and faces for emotion-based personalization. In: AfCAI"},{"issue":"1","key":"11271_CR144","doi-asserted-by":"crossref","first-page":"274","DOI":"10.1038\/s41597-022-01402-6","volume":"9","author":"K Kutt","year":"2022","unstructured":"Kutt K, Dr\u0105\u017cyk D, \u017buchowska L, Szel\u0105\u017cek M, Bobek S, Nalepa G (2022) Biraffe2, a multimodal dataset for emotion-based personalization in rich affective game environments. Sci Data 9(1):274","journal-title":"Sci Data"},{"key":"11271_CR145","doi-asserted-by":"crossref","unstructured":"Kye S, Moon J, Lee J, Choi I, Cheon D, Lee K (2017) Multimodal data collection framework for mental stress monitoring. In: Proceedings of the 2017 ACM international joint conference on pervasive and ubiquitous computing and proceedings of the 2017 ACM international symposium on wearable computers, pp 822\u2013829","DOI":"10.1145\/3123024.3125616"},{"issue":"1","key":"11271_CR146","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3702003","volume":"21","author":"G Lan","year":"2024","unstructured":"Lan G, Du Y, Yang Z (2024) Robust multimodal representation under uncertain missing modalities. ACM Trans Multimed Comput Commun Appl 21(1):1\u201323","journal-title":"ACM Trans Multimed Comput Commun Appl"},{"key":"11271_CR147","doi-asserted-by":"crossref","first-page":"443","DOI":"10.1016\/j.neucom.2021.05.103","volume":"470","author":"I Lauriola","year":"2022","unstructured":"Lauriola I, Lavelli A, Aiolli F (2022) An introduction to deep learning in natural language processing: models, techniques, and tools. Neurocomputing 470:443\u2013456","journal-title":"Neurocomputing"},{"key":"11271_CR148","doi-asserted-by":"crossref","unstructured":"Lefter I, Jonker CM, Tuente SK, Veling W, Bogaerts S (2017) Naa: A multimodal database of negative affect and aggression. In: Proc. IEEE seventh international conference on affective computing and intelligent interaction (ACII), pp 21\u201327","DOI":"10.1109\/ACII.2017.8273574"},{"issue":"3","key":"11271_CR149","doi-asserted-by":"crossref","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li S, Deng W (2022a) Deep facial expression recognition: a survey. IEEE Trans Affect Comput 13(3):1195\u20131215","journal-title":"IEEE Trans Affect Comput"},{"issue":"2","key":"11271_CR150","doi-asserted-by":"crossref","first-page":"881","DOI":"10.1109\/TAFFC.2020.2973158","volume":"13","author":"S Li","year":"2020","unstructured":"Li S, Deng W (2022b) A deeper look at facial expression dataset bias. IEEE Trans Affect Comput 13(2):881\u2013893","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR151","unstructured":"Li S, Tang H (2024) Multimodal alignment and fusion: A survey. arXiv preprint arXiv:2411.17040"},{"issue":"6","key":"11271_CR152","doi-asserted-by":"crossref","first-page":"913","DOI":"10.1007\/s12652-016-0406-z","volume":"8","author":"Y Li","year":"2017","unstructured":"Li Y, Tao J, Chao L, Bao W, Liu Y (2017) Cheavd: a chinese natural emotional audio-visual database. J Ambient Intell Humaniz Comput 8(6):913\u2013924","journal-title":"J Ambient Intell Humaniz Comput"},{"key":"11271_CR153","doi-asserted-by":"crossref","unstructured":"Li Y, Wang Y, Cui Z (2023) Decoupled multimodal distilling for emotion recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6631\u20136640","DOI":"10.1109\/CVPR52729.2023.00641"},{"key":"11271_CR154","unstructured":"Lian Z, Sun H, Sun L, Yi J, Liu B, Tao J (2024) Affectgpt: Dataset and framework for explainable multimodal emotion recognition. arXiv preprint arXiv:2407.07653"},{"issue":"2","key":"11271_CR155","doi-asserted-by":"crossref","first-page":"107","DOI":"10.1007\/s10055-024-00989-y","volume":"28","author":"R Lima","year":"2024","unstructured":"Lima R, Chirico A, Varandas R, Gamboa H, Gaggioli A, I Badia SB (2024) Multimodal emotion classification using machine learning in immersive and non-immersive virtual reality. Virtual Reality 28(2):107","journal-title":"Virtual Reality"},{"issue":"4","key":"11271_CR156","doi-asserted-by":"crossref","first-page":"2573","DOI":"10.3390\/app13042573","volume":"13","author":"W Lin","year":"2023","unstructured":"Lin W, Li C (2023) Review of studies on emotion recognition and judgment based on physiological signals. Appl Sci 13(4):2573","journal-title":"Appl Sci"},{"issue":"5","key":"11271_CR157","doi-asserted-by":"crossref","first-page":"e0196391","DOI":"10.1371\/journal.pone.0196391","volume":"13","author":"SR Livingstone","year":"2018","unstructured":"Livingstone SR, Russo FA (2018) The ryerson audio-visual database of emotional speech and song (ravdess): a dynamic, multimodal set of facial and vocal expressions in north american english. PLoS ONE 13(5):e0196391","journal-title":"PLoS ONE"},{"key":"11271_CR158","doi-asserted-by":"crossref","unstructured":"Lucey P, Cohn JF, Kanade T, Saragih J, Ambadar Z, Matthews I (2010) The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In: Proceedings of the IEEE computer society conference on computer vision and pattern Recognition-Workshops, pp 94\u2013101","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"11271_CR159","unstructured":"Lukowicz P, Pirkl G, Bannach D, Wagner F, Calatroni A, F\u00f6rster K, Holleczek T, Rossi M, Roggen D, Tr\u00f6ster G, et\u00a0al (2010) Recording a complex, multi modal activity data set for context recognition. In: 23th international conference on architecture of computing systems 2010, VDE, pp 1\u20136"},{"key":"11271_CR160","doi-asserted-by":"crossref","unstructured":"Ma F, Li Y, Xie Y, He Y, Zhang Y, Ren H, Liu Z, Yao W, Ren F, Yu FR, et\u00a0al (2024) A review of human emotion synthesis based on generative technology. arXiv preprint arXiv:2412.07116","DOI":"10.1109\/TAFFC.2025.3573878"},{"key":"11271_CR161","doi-asserted-by":"crossref","unstructured":"Ma K, Wang X, Yang X, Zhang M, Girard JM, Morency LP (2019) Elderreact: a multimodal dataset for recognizing emotional response in aging adults. In: 2019 international conference on multimodal interaction, pp 349\u2013357","DOI":"10.1145\/3340555.3353747"},{"key":"11271_CR162","doi-asserted-by":"crossref","unstructured":"Ma M, Ren J, Zhao L, Tulyakov S, Wu C, Peng X (2021) Smil: Multimodal learning with severely missing modality. Proceedings of the AAAI Conference on Artificial Intelligence 35:2302\u20132310","DOI":"10.1609\/aaai.v35i3.16330"},{"key":"11271_CR163","doi-asserted-by":"crossref","unstructured":"Mand AA, Wen JSJ, Sayeed MS, Swee SK (2017) Robust stress classifier using adaptive neuro-fuzzy classifier-linguistic hedges. 2017 International Conference on Robotics. IEEE, Automation and Sciences (ICORAS), pp 1\u20135","DOI":"10.1109\/ICORAS.2017.8308050"},{"key":"11271_CR164","doi-asserted-by":"crossref","unstructured":"Manolev J, Sullivan A, Slee R (2020) The datafication of discipline: Classdojo, surveillance and a performative classroom culture. In: The datafication of education, routledge, pp 37\u201352","DOI":"10.4324\/9780429341359-4"},{"issue":"1","key":"11271_CR165","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1057\/s41599-022-01483-z","volume":"10","author":"P Mantello","year":"2023","unstructured":"Mantello P, Ho MT, Nguyen MH, Vuong QH (2023) Machines that feel: behavioral determinants of attitude towards affect recognition technology-upgrading technology acceptance theory with the mindsponge model. Human Soc Sci Commun 10(1):1\u201316","journal-title":"Human Soc Sci Commun"},{"key":"11271_CR166","doi-asserted-by":"crossref","unstructured":"Mao X, Chen L, Fu L (2009) Multi-level speech emotion recognition based on hmm and ann. In: 2009 WRI World congress on computer science and information engineering. IEEE 7:225\u2013229","DOI":"10.1109\/CSIE.2009.113"},{"key":"11271_CR167","doi-asserted-by":"crossref","unstructured":"Martin O, Kotsia I, Macq B, Pitas I (2006) The enterface\u201905 audio-visual emotion database. In: Proceedings of the 22nd IEEE international Conference on data Engineering workshops (ICDEW\u201906), pp 8\u20138","DOI":"10.1109\/ICDEW.2006.145"},{"issue":"1","key":"11271_CR168","doi-asserted-by":"crossref","first-page":"1","DOI":"10.3758\/s13428-011-0124-6","volume":"44","author":"W Mason","year":"2012","unstructured":"Mason W, Suri S (2012) Conducting behavioral research on amazon\u2019s mechanical turk. Behav Res Methods 44(1):1\u201323","journal-title":"Behav Res Methods"},{"issue":"3","key":"11271_CR169","doi-asserted-by":"crossref","first-page":"276","DOI":"10.11613\/BM.2012.031","volume":"22","author":"ML McHugh","year":"2012","unstructured":"McHugh ML (2012) Interrater reliability: the kappa statistic. Biochemia Medica 22(3):276\u2013282","journal-title":"Biochemia Medica: Biochemia Medica"},{"key":"11271_CR170","doi-asserted-by":"crossref","unstructured":"McKeown G, Valstar MF, Cowie R, Pantic M (2010) The semaine corpus of emotionally coloured character interactions. In: Proceedings of the IEEE international conference on multimedia and expo (ICME), pp 1079\u20131084","DOI":"10.1109\/ICME.2010.5583006"},{"issue":"1","key":"11271_CR171","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1109\/T-AFFC.2011.20","volume":"3","author":"G McKeown","year":"2011","unstructured":"McKeown G, Valstar M, Cowie R, Pantic M, Schroder M (2011) The semaine database: annotated multimodal records of emotionally colored conversations between a person and a limited agent. IEEE Trans Affect Comput 3(1):5\u201317","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR172","doi-asserted-by":"crossref","first-page":"689","DOI":"10.1016\/j.procs.2020.07.101","volume":"175","author":"W Mellouk","year":"2020","unstructured":"Mellouk W, Handouzi W (2020) Facial emotion recognition using deep learning: review and insights. Procedia Comp Sci 175:689\u2013694","journal-title":"Procedia Comp Sci"},{"key":"11271_CR173","doi-asserted-by":"crossref","first-page":"108580","DOI":"10.1016\/j.knosys.2022.108580","volume":"244","author":"AI Middya","year":"2022","unstructured":"Middya AI, Nag B, Roy S (2022) Deep learning based multimodal emotion recognition using model-level fusion of audio-visual modalities. Knowl-Based Syst 244:108580","journal-title":"Knowl-Based Syst"},{"key":"11271_CR174","unstructured":"Miranda JA, Rituerto-Gonz\u00e1lez E, Guti\u00e9rrez-Mart\u00edn L, Luis-Mingueza C, Canabal MF, B\u00e1rcenas AR, Lanza-Guti\u00e9rrez JM, Pel\u00e1ez-Moreno C, L\u00f3pez-Ongil C (2022) Wemac: Women and emotion multi-modal affective computing dataset. arXiv preprint arXiv:2203.00456"},{"issue":"2","key":"11271_CR175","doi-asserted-by":"crossref","first-page":"479","DOI":"10.1109\/TAFFC.2018.2884461","volume":"12","author":"JA Miranda-Correa","year":"2018","unstructured":"Miranda-Correa JA, Abadi MK, Sebe N, Patras I (2018) Amigos: a dataset for affect, personality and mood research on individuals and groups. IEEE Trans Affect Comput 12(2):479\u2013493","journal-title":"IEEE Trans Affect Comput"},{"issue":"2","key":"11271_CR176","doi-asserted-by":"crossref","first-page":"239","DOI":"10.1162\/coli_a_00433","volume":"48","author":"SM Mohammad","year":"2022","unstructured":"Mohammad SM (2022) Ethics sheet for automatic emotion recognition and sentiment analysis. Comput Linguist 48(2):239\u2013278","journal-title":"Comput Linguist"},{"key":"11271_CR177","doi-asserted-by":"crossref","unstructured":"Morency LP, Mihalcea R, Doshi P (2011) Towards multimodal sentiment analysis: Harvesting opinions from the web. In: Proceedings of the 13th ACM international conference on multimodal interfaces, pp 169\u2013176","DOI":"10.1145\/2070481.2070509"},{"issue":"1","key":"11271_CR178","doi-asserted-by":"crossref","first-page":"39","DOI":"10.18178\/joig.6.1.39-43","volume":"6","author":"H Najadat","year":"2018","unstructured":"Najadat H, Abushaqra F (2018) Multimodal sentiment analysis of arabic videos. J Image and Grap 6(1):39\u201343","journal-title":"J Image and Grap"},{"key":"11271_CR179","doi-asserted-by":"crossref","unstructured":"Ni J, Li J, McAuley J (2019) Justifying recommendations using distantly-labeled reviews and fine-grained aspects. In: Proceedings of the conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 188\u2013197","DOI":"10.18653\/v1\/D19-1018"},{"key":"11271_CR180","doi-asserted-by":"crossref","unstructured":"Nojavanasghari B, Baltru\u0161aitis T, Hughes CE, Morency LP (2016) Emoreact: a multimodal approach and dataset for recognizing emotional responses in children. In: Proceedings of the 18th acm international conference on multimodal interaction, pp 137\u2013144","DOI":"10.1145\/2993148.2993168"},{"key":"11271_CR181","doi-asserted-by":"crossref","unstructured":"Otterdijk Mv, Neggers M, Torresen J, Barakova E (2024) Exploring human attribution of emotional intent to motion features in a humanoid robot. In: International conference on social robotics, Springer, pp 312\u2013323","DOI":"10.1007\/978-981-96-3519-1_29"},{"key":"11271_CR182","unstructured":"Pantic M, Valstar M, Rademaker R, Maat L (2005) Web-based database for facial expression analysis. In: IEEE international conference on multimedia and Expo, pp 5\u2013pp"},{"key":"11271_CR183","first-page":"330","volume":"7","author":"K P\u00e1pay","year":"2011","unstructured":"P\u00e1pay K, Szeghalmy S, Szekr\u00e9nyes I (2011) Hucomtech multimodal corpus annotation. Argumentum 7:330\u2013347","journal-title":"Argumentum"},{"issue":"1","key":"11271_CR184","doi-asserted-by":"crossref","first-page":"293","DOI":"10.1038\/s41597-020-00630-y","volume":"7","author":"CY Park","year":"2020","unstructured":"Park CY, Cha N, Kang S, Kim A, Khandoker AH, Hadjileontiadis L, Oh A, Jeong Y, Lee U (2020) K-emocon, a multimodal sensor dataset for continuous emotion recognition in naturalistic conversations. Sci Data 7(1):293","journal-title":"Sci Data"},{"issue":"32","key":"11271_CR185","doi-asserted-by":"crossref","first-page":"28663","DOI":"10.1007\/s12144-022-03932-z","volume":"42","author":"G Park","year":"2023","unstructured":"Park G, Chung J, Lee S (2023) Effect of ai chatbot emotional disclosure on user satisfaction and reuse intention for mental health counseling: a serial mediation model. Curr Psychol 42(32):28663\u201328673","journal-title":"Curr Psychol"},{"key":"11271_CR186","doi-asserted-by":"crossref","unstructured":"Park S, Shim HS, Chatterjee M, Sagae K, Morency LP (2014) Computational analysis of persuasiveness in social multimedia: A novel dataset and multimodal prediction approach. In: Proceedings of the 16th international conference on multimodal interaction, pp 50\u201357","DOI":"10.1145\/2663204.2663260"},{"key":"11271_CR187","doi-asserted-by":"crossref","unstructured":"Pereira M, P\u00e1dua F, Pereira A, Benevenuto F, Dalip D (2016) Fusing audio, textual, and visual features for sentiment analysis of news videos. In: Proceedings of the 10th international conference on web and social media, ICWSM, pp 659\u2013662","DOI":"10.1609\/icwsm.v10i1.14810"},{"issue":"16","key":"11271_CR188","doi-asserted-by":"crossref","first-page":"23783","DOI":"10.1007\/s11042-019-7691-4","volume":"78","author":"MH Pereira","year":"2019","unstructured":"Pereira MH, P\u00e1dua FL, Dalip DH, Benevenuto F, Pereira AC, Lacerda AM (2019) Multimodal approach for tension levels estimation in news videos. Multimed Tools and Appl 78(16):23783\u201323808","journal-title":"Multimed Tools and Appl"},{"key":"11271_CR189","doi-asserted-by":"crossref","unstructured":"Perepelkina O, Kazimirova E, Konstantinova M (2018) Ramas: Russian multimodal corpus of dyadic interaction for affective computing. In: International conference on speech and computer, Springer, pp 501\u2013510","DOI":"10.1007\/978-3-319-99579-3_52"},{"issue":"3","key":"11271_CR190","doi-asserted-by":"crossref","first-page":"38","DOI":"10.1109\/MIS.2013.9","volume":"28","author":"V P\u00e9rez Rosas","year":"2013","unstructured":"P\u00e9rez Rosas V, Mihalcea R, Morency LP (2013a) Multimodal sentiment analysis of spanish online videos. IEEE Intell Syst 28(3):38\u201345","journal-title":"IEEE Intell Syst"},{"key":"11271_CR191","unstructured":"P\u00e9rez-Rosas V, Mihalcea R, Morency LP (2013b) Utterance-level multimodal sentiment analysis. In: Association for computational linguistics (ACL), pp 973\u2013982"},{"key":"11271_CR192","doi-asserted-by":"crossref","unstructured":"Petersen IT (2024) Principles of psychological assessment: With applied examples in R. CRC Press","DOI":"10.1201\/9781003357421"},{"key":"11271_CR193","unstructured":"Petrik S (2004) Wizard of oz experiments on speech dialogue systems. Design and Realisation with a New Integrated Simulation Environment Masters Graz University of Technology, Graz Institute of Signal Processing and Speech Communication"},{"key":"11271_CR194","doi-asserted-by":"crossref","first-page":"149","DOI":"10.1016\/j.inffus.2022.10.004","volume":"91","author":"HT Phan","year":"2023","unstructured":"Phan HT, Nguyen NT, Hwang D (2023) Aspect-level sentiment analysis: a survey of graph convolutional network methods. Inf Fusion 91:149\u2013172","journal-title":"Inf Fusion"},{"key":"11271_CR195","unstructured":"Pichora-Fuller MK, Dupuis K (2020) Toronto emotional speech set (TESS)"},{"key":"11271_CR196","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1016\/B978-0-12-558701-3.50007-7","volume":"1","author":"R Plutchik","year":"1980","unstructured":"Plutchik R (1980) A general psychoevolutionary theory of emotion. Theories of Emotion 1:3\u201331","journal-title":"Theories of Emotion"},{"key":"11271_CR197","doi-asserted-by":"crossref","unstructured":"Pontiki M, Galanis D, Papageorgiou H, Manandhar S, Androutsopoulos I (2015) Semeval-2015 task 12: Aspect based sentiment analysis. In: Proceedings of the 9th international workshop on semantic evaluation (SemEval 2015), pp 486\u2013495","DOI":"10.18653\/v1\/S15-2082"},{"key":"11271_CR198","doi-asserted-by":"crossref","unstructured":"Pontiki M, Galanis D, Papageorgiou H, Androutsopoulos I, Manandhar S, AL-Smadi M, Al-Ayyoub M, Zhao Y, Qin B, De\u00a0Clercq O, et\u00a0al (2016) Semeval-2016 task 5: Aspect based sentiment analysis. In: ProWorkshop on Semantic Evaluation (SemEval-2016), Association for computational linguistics, pp 19\u201330","DOI":"10.18653\/v1\/S16-1002"},{"key":"11271_CR199","doi-asserted-by":"crossref","first-page":"98","DOI":"10.1016\/j.inffus.2017.02.003","volume":"37","author":"S Poria","year":"2017","unstructured":"Poria S, Cambria E, Bajpai R, Hussain A (2017) A review of affective computing: from unimodal analysis to multimodal fusion. Inf Fusion 37:98\u2013125","journal-title":"Inf Fusion"},{"key":"11271_CR200","doi-asserted-by":"crossref","unstructured":"Poria S, Hazarika D, Majumder N, Naik G, Cambria E, Mihalcea R (2018) Meld: A multimodal multi-party dataset for emotion recognition in conversations. arXiv preprint arXiv:1810.02508","DOI":"10.18653\/v1\/P19-1050"},{"key":"11271_CR201","doi-asserted-by":"crossref","first-page":"100943","DOI":"10.1109\/ACCESS.2019.2929050","volume":"7","author":"S Poria","year":"2019","unstructured":"Poria S, Majumder N, Mihalcea R, Hovy E (2019) Emotion recognition in conversation: research challenges, datasets, and recent advances. IEEE Access 7:100943\u2013100953","journal-title":"IEEE Access"},{"key":"11271_CR202","unstructured":"Povey D, Ghoshal A, Boulianne G, Burget L, Glembek O, Goel N, Hannemann M, Motlicek P, Qian Y, Schwarz P, et\u00a0al (2011) The Kaldi speech recognition toolkit. In: Proceedings of the IEEE workshop on automatic speech recognition and understanding"},{"key":"11271_CR203","unstructured":"Pustejovsky J, Stubbs A (2012) Natural Language Annotation for Machine Learning: A guide to corpus-building for applications. O\u2019Reilly Media, Inc."},{"issue":"2","key":"11271_CR204","first-page":"13","volume":"81","author":"R Pusztahelyi","year":"2020","unstructured":"Pusztahelyi R et al (2020) Emotional ai and its challenges in the viewpoint of online marketing. Curentul Juridic 81(2):13\u201331","journal-title":"Curentul Juridic"},{"key":"11271_CR205","doi-asserted-by":"crossref","unstructured":"Rabbimov I, Mporas I, Simaki V, Kobilov S (2020) Investigating the effect of emoji in opinion classification of uzbek movie review comments. In: International conference on speech and computer, Springer, pp 435\u2013445","DOI":"10.1007\/978-3-030-60276-5_42"},{"key":"11271_CR206","doi-asserted-by":"crossref","first-page":"203","DOI":"10.1016\/j.inffus.2021.12.003","volume":"81","author":"A Rahate","year":"2022","unstructured":"Rahate A, Walambe R, Ramanna S, Kotecha K (2022) Multimodal co-learning: challenges, applications with datasets, recent advances and future directions. Inf Fusion 81:203\u2013239","journal-title":"Inf Fusion"},{"key":"11271_CR207","doi-asserted-by":"crossref","unstructured":"Rathi T, Tripathy M (2024) Analyzing the influence of different speech data corpora and speech features on speech emotion recognition: A review. Speech Communication p 103102","DOI":"10.1016\/j.specom.2024.103102"},{"issue":"3","key":"11271_CR208","doi-asserted-by":"crossref","first-page":"252","DOI":"10.1109\/34.75512","volume":"13","author":"SJ Raudys","year":"1991","unstructured":"Raudys SJ, Jain AK et al (1991) Small sample size effects in statistical pattern recognition: recommendations for practitioners. IEEE Trans Pattern Anal Mach Intell 13(3):252\u2013264","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11271_CR209","doi-asserted-by":"crossref","unstructured":"Ringeval F, Sonderegger A, Sauer J, Lalanne D (2013) Introducing the recola multimodal corpus of remote collaborative and affective interactions. In: 10th IEEE international conference and workshops on automatic face and gesture recognition (FG), pp 1\u20138","DOI":"10.1109\/FG.2013.6553805"},{"key":"11271_CR210","unstructured":"Rizvi SSA, Kumar S, Seth A, Narang P (2025) Affectsrnet: Facial emotion-aware super-resolution network. arXiv preprint arXiv:2502.09932"},{"issue":"5","key":"11271_CR211","doi-asserted-by":"crossref","first-page":"1081","DOI":"10.1109\/TNSRE.2020.2980223","volume":"28","author":"KA Robbins","year":"2020","unstructured":"Robbins KA, Touryan J, Mullen T, Kothe C, Bigdely-Shamlo N (2020) How sensitive are eeg results to preprocessing methods: a benchmarking study. IEEE Trans Neural Syst Rehabil Eng 28(5):1081\u20131090","journal-title":"IEEE Trans Neural Syst Rehabil Eng"},{"key":"11271_CR212","doi-asserted-by":"crossref","unstructured":"Roshanaei M (2024) Towards best practices for mitigating artificial intelligence implicit bias in shaping diversity, inclusion and equity in higher education. Education and Information Technologies pp 1\u201326","DOI":"10.1007\/s10639-024-12605-2"},{"key":"11271_CR213","unstructured":"Rozgi\u0107 V, Ananthakrishnan S, Saleem S, Kumar R, Prasad R (2012) Ensemble of SVM trees for multimodal emotion recognition. In: Proceedings of the IEEE signal & information processing association annual summit and conference (APSIPA ASC), Asia-Pacific, pp 1\u20134"},{"key":"11271_CR214","doi-asserted-by":"crossref","unstructured":"Ruotsalo T, Traver VJ, Kawala-Sterniuk A, Leiva LA (2024) Affective relevance. IEEE Intelligent Systems","DOI":"10.1109\/MIS.2024.3391508"},{"issue":"6","key":"11271_CR215","doi-asserted-by":"crossref","first-page":"1161","DOI":"10.1037\/h0077714","volume":"39","author":"JA Russell","year":"1980","unstructured":"Russell JA (1980) A circumplex model of affect. J Pers Soc Psychol 39(6):1161","journal-title":"J Pers Soc Psychol"},{"key":"11271_CR216","doi-asserted-by":"crossref","unstructured":"Sapi\u0144ski T, Kami\u0144ska D, Pelikant A, Ozcinar C, Avots E, Anbarjafari G (2018) Multimodal database of emotional speech, video and gestures. In: International conference on pattern recognition, Springer, pp 153\u2013163","DOI":"10.1007\/978-3-030-05792-3_15"},{"issue":"1","key":"11271_CR217","first-page":"53","volume":"2","author":"A Saxena","year":"2020","unstructured":"Saxena A, Khanna A, Gupta D (2020) Emotion recognition and detection methods: a comprehensive survey. J Artif Intell Syst 2(1):53\u201379","journal-title":"J Artif Intell Syst"},{"key":"11271_CR218","doi-asserted-by":"crossref","first-page":"105450","DOI":"10.1016\/j.neubiorev.2023.105450","volume":"158","author":"D Schiller","year":"2024","unstructured":"Schiller D, Alessandra N, Alia-Klein N, Becker S, Cromwell HC, Dolcos F, Eslinger PJ, Frewen P, Kemp AH, Pace-Schott EF et al (2024) The human affectome. Neurosci Biobehav Rev 158:105450","journal-title":"Neurosci Biobehav Rev"},{"key":"11271_CR219","unstructured":"Schimmel A, Doyran M, Akdag A, Bakkes S, Kaya H, Poppe R, Salah A (2019) Multi-person board game affect analysis dataset: Mp-bgaad. In: Proceedings eNTERFACE\u201919, computer engineering department bilkent university"},{"key":"11271_CR220","doi-asserted-by":"crossref","unstructured":"Schmidt P, Reiss A, Duerichen R, Marberger C, Van\u00a0Laerhoven K (2018) Introducing wesad, a multimodal dataset for wearable stress and affect detection. In: Proceedings of the 20th ACM international conference on multimodal interaction, pp 400\u2013408","DOI":"10.1145\/3242969.3242985"},{"key":"11271_CR221","doi-asserted-by":"crossref","unstructured":"Schuller B, Reiter S, Muller R, Al-Hames M, Lang M, Rigoll G (2005) Speaker independent speech emotion recognition by ensemble classification. In: IEEE international conference on multimedia and expo, pp 864\u2013867","DOI":"10.1109\/ICME.2005.1521560"},{"key":"11271_CR222","doi-asserted-by":"crossref","unstructured":"Serban IV, Lowe R, Henderson P, Charlin L, Pineau J (2018) A survey of available corpora for building data-driven dialogue systems: The journal version. Dialogue & Discourse 9(1):1\u201349","DOI":"10.5087\/dad.2018.101"},{"key":"11271_CR223","doi-asserted-by":"crossref","unstructured":"Shahriar S, Allana S, Fard MH, Dara R (2023) A survey of privacy risks and mitigation strategies in the artificial intelligence life cycle. IEEE Access","DOI":"10.1109\/ACCESS.2023.3287195"},{"key":"11271_CR224","doi-asserted-by":"crossref","unstructured":"Sharma C, Bhageria D, Scott W, Pykl S, Das A, Chakraborty T, Pulabaigari V, Gamback B (2020) Semeval-2020 task 8: Memotion analysis\u2013the visuo-lingual metaphor! arXiv preprint arXiv:2008.03781","DOI":"10.18653\/v1\/2020.semeval-1.99"},{"key":"11271_CR225","doi-asserted-by":"crossref","first-page":"548","DOI":"10.33314\/jnhrc.v17i4.1042","volume":"17","author":"B Shrestha","year":"2019","unstructured":"Shrestha B, Dunn L (2019) The declaration of helsinki on medical research involving human subjects: a review of seventh revision. J Nepal Health Res Counc 17:548\u2013552","journal-title":"J Nepal Health Res Counc"},{"issue":"6","key":"11271_CR226","first-page":"47","volume":"6","author":"MFH Siddiqui","year":"2022","unstructured":"Siddiqui MFH, Dhakal P, Yang X, Javaid AY (2022) A survey on databases for multimodal emotion recognition and an introduction to the viri (visible and infrared image) database. Multimodal Tech Inter 6(6):47","journal-title":"Multimodal Tech Inter"},{"key":"11271_CR227","doi-asserted-by":"crossref","first-page":"17","DOI":"10.1007\/s12193-013-0129-9","volume":"8","author":"I Siegert","year":"2014","unstructured":"Siegert I, B\u00f6ck R, Wendemuth A (2014) Inter-rater reliability for emotion annotation in human-computer interaction: comparison and methodological improvements. J Multimod User Interfaces 8:17\u201328","journal-title":"J Multimod User Interfaces"},{"issue":"1","key":"11271_CR228","doi-asserted-by":"crossref","first-page":"42","DOI":"10.1109\/T-AFFC.2011.25","volume":"3","author":"M Soleymani","year":"2012","unstructured":"Soleymani M, Lichtenauer J, Pun T, Pantic M (2012) A multimodal database for affect recognition and implicit tagging. IEEE Trans Affect Comput 3(1):42\u201355","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR229","doi-asserted-by":"crossref","first-page":"103967","DOI":"10.1016\/j.apergo.2023.103967","volume":"109","author":"Y Song","year":"2023","unstructured":"Song Y, Tao D, Luximon Y (2023) In robot we trust? the effect of emotional expressions and contextual cues on anthropomorphic trustworthiness. Appl Ergon 109:103967","journal-title":"Appl Ergon"},{"issue":"2","key":"11271_CR230","doi-asserted-by":"crossref","first-page":"147","DOI":"10.1109\/TAFFC.2016.2625250","volume":"9","author":"R Subramanian","year":"2016","unstructured":"Subramanian R, Wache J, Abadi MK, Vieriu RL, Winkler S, Sebe N (2016) Ascertain: emotion and personality recognition using commercial sensors. IEEE Trans Affect Comput 9(2):147\u2013160","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR231","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1007\/s10772-018-9491-z","volume":"21","author":"M Swain","year":"2018","unstructured":"Swain M, Routray A, Kabisatpathy P (2018) Databases, features and classifiers for speech emotion recognition: a review. Int J Speech Technol 21:93\u2013120","journal-title":"Int J Speech Technol"},{"issue":"5","key":"11271_CR232","doi-asserted-by":"crossref","first-page":"1167","DOI":"10.28991\/ESJ-2022-06-05-017","volume":"6","author":"K Tawsif","year":"2022","unstructured":"Tawsif K, Aziz NAA, Raja JE, Hossen J, Jesmeen M (2022) A systematic review on emotion recognition system using physiological signals: data acquisition and methodology. Emerg Sci J 6(5):1167\u20131198","journal-title":"Emerg Sci J"},{"key":"11271_CR233","unstructured":"Tekalp AM (2015) Digital video processing. Prentice Hall Press"},{"key":"11271_CR234","doi-asserted-by":"crossref","first-page":"24976","DOI":"10.1109\/ACCESS.2022.3153787","volume":"10","author":"MT Teye","year":"2022","unstructured":"Teye MT, Missah YM, Ahene E, Frimpong T (2022) Evaluation of conversational agents: understanding culture, context and environment in emotion detection. IEEE Access 10:24976\u201324984","journal-title":"IEEE Access"},{"key":"11271_CR235","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.102075","volume":"103","author":"S Umirzakova","year":"2024","unstructured":"Umirzakova S, Ahmad S, Khan LU, Whangbo T (2024) Medical image super-resolution for smart healthcare applications: a comprehensive survey. Inf Fusion 103:102075","journal-title":"Inf Fusion"},{"issue":"2","key":"11271_CR236","first-page":"233","volume":"27","author":"L Urquhart","year":"2022","unstructured":"Urquhart L, Miranda D, Podoletz L (2022) Policing the smart home: the internet of things as \u2018invisible witnesses\u2019. Inf Polity 27(2):233\u2013246","journal-title":"Inf Polity"},{"issue":"1","key":"11271_CR237","first-page":"100165","volume":"3","author":"P Varsha","year":"2023","unstructured":"Varsha P (2023) How can we manage biases in artificial intelligence systems-a systematic literature review. Int J Inf Manag Data Insights 3(1):100165","journal-title":"Int J Inf Manag Data Insights"},{"key":"11271_CR238","doi-asserted-by":"crossref","unstructured":"Velana M, Gruss S, Layher G, Thiam P, Zhang Y, Schork D, Kessler V, Meudt S, Neumann H, Kim J, et\u00a0al (2017) The senseemotion database: A multimodal database for the development and systematic validation of an automatic pain-and emotion-recognition system. In: Multimodal pattern recognition of social signals in Human-Computer-Interaction: 4th IAPR TC 9 Workshop, MPRSS 2016, Cancun, Mexico, December 4, 2016, Revised Selected Papers 4, Springer, pp 127\u2013139","DOI":"10.1007\/978-3-319-59259-6_11"},{"key":"11271_CR239","doi-asserted-by":"crossref","first-page":"1387089","DOI":"10.3389\/fpsyg.2024.1387089","volume":"15","author":"AOR Vistorte","year":"2024","unstructured":"Vistorte AOR, Deroncele-Acosta A, Ayala JLM, Barrasa A, L\u00f3pez-Granero C, Mart\u00ed-Gonz\u00e1lez M (2024) Integrating artificial intelligence to assess emotions in learning environments: a systematic literature review. Front Psychol 15:1387089","journal-title":"Front Psychol"},{"key":"11271_CR240","doi-asserted-by":"crossref","unstructured":"Wang J, Sun C, Li S, Liu X, Si L, Zhang M, Zhou G (2019) Aspect sentiment classification towards question-answering with reinforced bidirectional attention network. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 3548\u20133557","DOI":"10.18653\/v1\/P19-1345"},{"key":"11271_CR241","unstructured":"Wang K, Zhang Q, Liao S (2014a) A database of elderly emotional speech. In: Proc. Int. Symp. Signal Process. Biomed. Eng Informat, pp 549\u2013553"},{"key":"11271_CR242","doi-asserted-by":"crossref","unstructured":"Wang M, Cao D, Li L, Li S, Ji R (2014b) Microblog sentiment analysis based on cross-media bag-of-words model. In: Proceedings of international conference on internet multimedia computing and service, ACM, p\u00a076","DOI":"10.1145\/2632856.2632912"},{"issue":"4","key":"11271_CR243","doi-asserted-by":"crossref","first-page":"39","DOI":"10.1007\/s13735-024-00347-3","volume":"13","author":"R Wang","year":"2024","unstructured":"Wang R, Zhu J, Wang S, Wang T, Huang J, Zhu X (2024) Multi-modal emotion recognition using tensor decomposition fusion and self-supervised multi-tasking. Int J Multimed Inf Retriev 13(4):39","journal-title":"Int J Multimed Inf Retriev"},{"key":"11271_CR244","doi-asserted-by":"crossref","first-page":"19","DOI":"10.1016\/j.inffus.2022.03.009","volume":"83","author":"Y Wang","year":"2022","unstructured":"Wang Y, Song W, Tao W, Liotta A, Yang D, Li X, Gao S, Sun Y, Ge W, Zhang W et al (2022) A systematic review on affective computing: emotion models, databases, and recent advances. Inf Fusion 83:19\u201352","journal-title":"Inf Fusion"},{"key":"11271_CR245","doi-asserted-by":"crossref","unstructured":"Wang Y, Yu H, Gao W, Xia Y, Nduka C (2023) Mgeed: A multimodal genuine emotion and expression detection database. IEEE Trans on Affect Comput 15(2): 606-619","DOI":"10.1109\/TAFFC.2023.3286351"},{"issue":"7","key":"11271_CR246","doi-asserted-by":"crossref","first-page":"5731","DOI":"10.1007\/s10462-022-10144-1","volume":"55","author":"M Wankhade","year":"2022","unstructured":"Wankhade M, Rao ACS, Kulkarni C (2022) A survey on sentiment analysis methods, applications, and challenges. Artif Intell Rev 55(7):5731\u20135780","journal-title":"Artif Intell Rev"},{"key":"11271_CR247","unstructured":"Weinbaum C, Landree E, Blumenthal MS, Piquado T, Gutierrez CI (2019) Ethics in scientific research. RAND Corporation"},{"issue":"4","key":"11271_CR248","doi-asserted-by":"crossref","first-page":"472","DOI":"10.24908\/ss.v22i4.18327","volume":"22","author":"T Wiehn","year":"2024","unstructured":"Wiehn T (2024) Synthetic data: from data scarcity to data pollution. Surveill Soc 22(4):472\u2013476","journal-title":"Surveill Soc"},{"issue":"3","key":"11271_CR249","doi-asserted-by":"crossref","first-page":"46","DOI":"10.1109\/MIS.2013.34","volume":"28","author":"M W\u00f6llmer","year":"2013","unstructured":"W\u00f6llmer M, Weninger F, Knaup T, Schuller B, Sun C, Sagae K, Morency LP (2013) Youtube movie reviews: sentiment analysis in an audio-visual context. IEEE Intell Syst 28(3):46\u201353","journal-title":"IEEE Intell Syst"},{"key":"11271_CR250","volume-title":"Digital image processing","author":"RE Woods","year":"2019","unstructured":"Woods RE, Gonzalez RC (2019) Digital image processing. Fourth edition, Pearson"},{"issue":"1","key":"11271_CR251","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1109\/T-AFFC.2010.16","volume":"2","author":"CH Wu","year":"2011","unstructured":"Wu CH, Liang WB (2011) Emotion recognition of affective speech based on multiple classifiers using acoustic-prosodic information and semantic labels. IEEE Trans Affect Comput 2(1):10\u201321","journal-title":"IEEE Trans Affect Comput"},{"key":"11271_CR252","unstructured":"Wu R, Wang H, Chen HT, Carneiro G (2024) Deep multimodal learning with missing modality: A survey. arXiv preprint arXiv:2409.07825"},{"key":"11271_CR253","doi-asserted-by":"crossref","unstructured":"Xing X, Jin Z, Jin D, Wang B, Zhang Q, Huang X (2020) Tasty burgers, soggy fries: Probing aspect robustness in aspect-based sentiment analysis. In: Proceedings of the 2020 conference on empirical methods in natural language processing, p 3594-3605","DOI":"10.18653\/v1\/2020.emnlp-main.292"},{"key":"11271_CR254","doi-asserted-by":"crossref","unstructured":"Xu H, Zhang H, Han K, Wang Y, Peng Y, Li X (2019a) Learning alignment for multimodal emotion recognition from speech. arXiv preprint arXiv:1909.05645","DOI":"10.21437\/Interspeech.2019-3247"},{"key":"11271_CR255","doi-asserted-by":"crossref","unstructured":"Xu N, Mao W, Chen G (2019) Multi-interactive memory network for aspect based multimodal sentiment analysis. Proceedings of the AAAI Conference on Artificial Intelligence 33:371\u2013378","DOI":"10.1609\/aaai.v33i01.3301371"},{"issue":"10","key":"11271_CR256","doi-asserted-by":"crossref","first-page":"12113","DOI":"10.1109\/TPAMI.2023.3275156","volume":"45","author":"P Xu","year":"2023","unstructured":"Xu P, Zhu X, Clifton DA (2023) Multimodal learning with transformers: a survey. IEEE Trans Pattern Anal Mach Intell 45(10):12113\u201312132","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11271_CR257","doi-asserted-by":"crossref","unstructured":"Xu P, Shao W, Zhang K, Gao P, Liu S, Lei M, Meng F, Huang S, Qiao Y, Luo P (2024) Lvlm-ehub: A comprehensive evaluation benchmark for large vision-language models. IEEE Trans on Pattern Anal Mach Intell","DOI":"10.1109\/TPAMI.2024.3507000"},{"key":"11271_CR258","doi-asserted-by":"crossref","unstructured":"Xu T, White J, Kalkan S, Gunes H (2020) Investigating bias and fairness in facial expression recognition. In: Computer vision\u2013ECCV 2020 workshops: Glasgow, UK, August 23\u201328, 2020, Proceedings, Part VI 16, Springer, pp 506\u2013523","DOI":"10.1007\/978-3-030-65414-6_35"},{"key":"11271_CR259","doi-asserted-by":"crossref","unstructured":"Xue Z, Marculescu R (2023) Dynamic multimodal fusion. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2575\u20132584","DOI":"10.1109\/CVPRW59228.2023.00256"},{"key":"11271_CR260","doi-asserted-by":"crossref","unstructured":"You Q, Luo J, Jin H, Yang J (2015) Robust image sentiment analysis using progressively trained and domain transferred deep networks. arXiv preprint arXiv:1509.06041","DOI":"10.1609\/aaai.v29i1.9179"},{"key":"11271_CR261","doi-asserted-by":"crossref","unstructured":"You Q, Luo J, Jin H, Yang J (2016) Cross-modality consistent regression for joint visual-textual sentiment analysis of social multimedia. In: Proceedings of the ninth ACM international conference on web search and data mining, pp 13\u201322","DOI":"10.1145\/2835776.2835779"},{"key":"11271_CR262","doi-asserted-by":"crossref","unstructured":"Yu J, Jiang J (2019) Adapting bert for target-oriented multimodal sentiment classification. In: Proceedings of the 28th international Joint conference on artifici al intelligence (IJCAI-19)","DOI":"10.24963\/ijcai.2019\/751"},{"key":"11271_CR263","doi-asserted-by":"crossref","unstructured":"Zachar P, Ellis RD (eds) (2012) Categorical versus dimensional models of affect. John Benjamins Publishing Company","DOI":"10.1075\/ceb.7"},{"issue":"6","key":"11271_CR264","doi-asserted-by":"crossref","first-page":"82","DOI":"10.1109\/MIS.2016.94","volume":"31","author":"A Zadeh","year":"2016","unstructured":"Zadeh A, Zellers R, Pincus E, Morency LP (2016) Multimodal sentiment intensity analysis in videos: facial gestures and verbal messages. IEEE Intell Syst 31(6):82\u201388","journal-title":"IEEE Intell Syst"},{"key":"11271_CR265","doi-asserted-by":"crossref","unstructured":"Zadeh AB, Liang PP, Poria S, Cambria E, Morency LP (2018) Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In: Proceedings of the 56th annual meeting of the association for computational linguistics (Volume 1: Long Papers), pp 2236\u20132246","DOI":"10.18653\/v1\/P18-1208"},{"key":"11271_CR266","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.cviu.2015.03.015","volume":"138","author":"S Zafeiriou","year":"2015","unstructured":"Zafeiriou S, Zhang C, Zhang Z (2015) A survey on face detection in the wild: past, present and future. Comput Vis Image Underst 138:1\u201324","journal-title":"Comput Vis Image Underst"},{"issue":"3","key":"11271_CR267","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3388790","volume":"53","author":"S Zepf","year":"2020","unstructured":"Zepf S, Hernandez J, Schmitt A, Minker W, Picard RW (2020) Driver emotion recognition for intelligent vehicles: a survey. ACM Comput Surv (CSUR) 53(3):1\u201330","journal-title":"ACM Comput Surv (CSUR)"},{"issue":"2","key":"11271_CR268","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3309543","volume":"37","author":"X Zhan","year":"2019","unstructured":"Zhan X, Wang Y, Rao Y, Li Q (2019) Learning from multi-annotator data: a noise-aware classification framework. ACM Trans on Inf Syst (TOIS) 37(2):1\u201328","journal-title":"ACM Trans on Inf Syst (TOIS)"},{"key":"11271_CR269","unstructured":"Zhang J, Jia H (2008) Design of speech corpus for mandarin text to speech. In: The blizzard challenge 2008 workshop"},{"key":"11271_CR270","volume":"100","author":"J Zhang","year":"2024","unstructured":"Zhang J, Chen Q, Lu J, Wang X, Liu L, Feng Y (2024) Emotional expression by artificial intelligence chatbots to improve customer satisfaction: underlying mechanism and boundary conditions. Tour Manage 100:104835","journal-title":"Tour Manage"},{"issue":"8","key":"11271_CR271","doi-asserted-by":"crossref","first-page":"1819","DOI":"10.1109\/TKDE.2013.39","volume":"26","author":"ML Zhang","year":"2013","unstructured":"Zhang ML, Zhou ZH (2013) A review on multi-label learning algorithms. IEEE Trans Knowl Data Eng 26(8):1819\u20131837","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"11271_CR272","unstructured":"Zhang Q, Wei Y, Han Z, Fu H, Peng X, Deng C, Hu Q, Xu C, Wen J, Hu D, et\u00a0al (2024b) Multimodal fusion on low-quality data: A comprehensive survey. arXiv preprint arXiv:2404.18947"},{"key":"11271_CR273","doi-asserted-by":"crossref","unstructured":"Zhao G, Li Y, Xu Q (2022a) From emotion ai to cognitive ai. International journal of network dynamics and intelligence pp 65\u201372","DOI":"10.53941\/ijndi0101006"},{"key":"11271_CR274","doi-asserted-by":"crossref","unstructured":"Zhao J, Zhang T, Hu J, Liu Y, Jin Q, Wang X, Li H (2022b) M3ed: Multi-modal multi-scene multi-label emotional dialogue database. arXiv preprint arXiv:2205.10237","DOI":"10.18653\/v1\/2022.acl-long.391"},{"issue":"3","key":"11271_CR275","doi-asserted-by":"crossref","first-page":"1252","DOI":"10.1080\/10494820.2020.1826985","volume":"31","author":"J Zhou","year":"2023","unstructured":"Zhou J, Jm Ye (2023) Sentiment analysis in education research: a review of journal publications. Interact Learn Environ 31(3):1252\u20131264","journal-title":"Interact Learn Environ"},{"key":"11271_CR276","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.specom.2021.11.006","volume":"137","author":"K Zhou","year":"2022","unstructured":"Zhou K, Sisman B, Liu R, Li H (2022) Emotional voice conversion: theory, databases and esd. Speech Commun 137:1\u201318","journal-title":"Speech Commun"},{"issue":"2","key":"11271_CR277","doi-asserted-by":"crossref","first-page":"1253","DOI":"10.1007\/s10586-022-03705-0","volume":"26","author":"Z Zhou","year":"2023","unstructured":"Zhou Z, Asghar MA, Nazir D, Siddique K, Shorfuzzaman M, Mehmood RM (2023) An ai-empowered affect recognition model for healthcare and emotional well-being using physiological signals. Clust Comput 26(2):1253\u20131266","journal-title":"Clust Comput"},{"issue":"1","key":"11271_CR278","doi-asserted-by":"crossref","first-page":"44","DOI":"10.1093\/nsr\/nwx106","volume":"5","author":"ZH Zhou","year":"2018","unstructured":"Zhou ZH (2018) A brief introduction to weakly supervised learning. Natl Sci Rev 5(1):44\u201353","journal-title":"Natl Sci Rev"},{"key":"11271_CR279","doi-asserted-by":"crossref","first-page":"306","DOI":"10.1016\/j.inffus.2023.02.028","volume":"95","author":"L Zhu","year":"2023","unstructured":"Zhu L, Zhu Z, Zhang C, Xu Y, Kong X (2023) Multimodal sentiment analysis based on fusion methods: a survey. Inf Fusion 95:306\u2013325","journal-title":"Inf Fusion"},{"issue":"18","key":"11271_CR280","doi-asserted-by":"crossref","first-page":"56039","DOI":"10.1007\/s11042-023-17347-w","volume":"83","author":"X Zhu","year":"2024","unstructured":"Zhu X, Huang Y, Wang X, Wang R (2024) Emotion recognition based on brain-like multimodal hierarchical perception. Multimed Tools and Appl 83(18):56039\u201356057","journal-title":"Multimed Tools and Appl"}],"container-title":["Artificial Intelligence Review"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10462-025-11271-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10462-025-11271-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10462-025-11271-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,12]],"date-time":"2025-09-12T18:10:42Z","timestamp":1757700642000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10462-025-11271-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,13]]},"references-count":277,"journal-issue":{"issue":"10","published-online":{"date-parts":[[2025,10]]}},"alternative-id":["11271"],"URL":"https:\/\/doi.org\/10.1007\/s10462-025-11271-1","relation":{},"ISSN":["1573-7462"],"issn-type":[{"value":"1573-7462","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8,13]]},"assertion":[{"value":"20 May 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 August 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"334"}}