{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T16:47:57Z","timestamp":1776876477761,"version":"3.51.2"},"reference-count":92,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T00:00:00Z","timestamp":1773360000000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Applied Soft Computing"],"published-print":{"date-parts":[[2026,7]]},"DOI":"10.1016\/j.asoc.2026.114998","type":"journal-article","created":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T07:51:33Z","timestamp":1773474693000},"page":"114998","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["A deep learning feature mapping algorithm for emotion detection via facial and audio signals"],"prefix":"10.1016","volume":"197","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5999-2134","authenticated-orcid":false,"given":"Mohammad Hassan","family":"Tayarani Najaran","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-9174-5044","authenticated-orcid":false,"given":"Shamim Ibne","family":"Shahid","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1797-682X","authenticated-orcid":false,"given":"Frank","family":"Foerster","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0186-3580","authenticated-orcid":false,"given":"Volker","family":"Steuber","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.asoc.2026.114998_bib0005","doi-asserted-by":"crossref","first-page":"123","DOI":"10.1007\/s11412-016-9232-8","article-title":"Group emotions: the social and cognitive functions of emotions in argumentation","volume":"11","author":"Polo","year":"2016","journal-title":"Int. J. Comput.-Support. Collab. Learn."},{"key":"10.1016\/j.asoc.2026.114998_bib0010","doi-asserted-by":"crossref","unstructured":"L. Perlovsky, Aesthetic emotions, what are their cognitive functions? 2014.","DOI":"10.3389\/fpsyg.2014.00098"},{"key":"10.1016\/j.asoc.2026.114998_bib0015","series-title":"Handbook of Communication and Emotion","first-page":"49","article-title":"Principles of communication and emotion in social interaction","author":"Andersen","year":"1996"},{"key":"10.1016\/j.asoc.2026.114998_bib0020","doi-asserted-by":"crossref","DOI":"10.1016\/j.beproc.2019.06.004","article-title":"A place for emotions in behavior systems research","volume":"166","author":"Burghardt","year":"2019","journal-title":"Behav. Processes"},{"key":"10.1016\/j.asoc.2026.114998_bib0025","series-title":"Emotional Well-Being in Educational Policy and Practice","first-page":"81","article-title":"From emotional and psychological well-being to character education: challenging policy discourses of behavioural science and \u2018vulnerability\u2019","author":"Ecclestone","year":"2017"},{"key":"10.1016\/j.asoc.2026.114998_bib0030","series-title":"Verbal and Nonverbal Communication Behaviours: COST Action 2102 International Workshop, Vietri Sul Mare, Italy, March 29\u201331, 2007, Revised Selected and Invited Papers","first-page":"74","article-title":"Meaningful parameters in emotion characterisation","author":"Navas","year":"2007"},{"key":"10.1016\/j.asoc.2026.114998_bib0035","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1111\/j.1475-6811.1996.tb00105.x","article-title":"Is love a \u201cbasic\u201d emotion?","volume":"3","author":"Shaver","year":"1996","journal-title":"Pers. Relat."},{"key":"10.1016\/j.asoc.2026.114998_bib0040","doi-asserted-by":"crossref","first-page":"1191","DOI":"10.1126\/science.1076358","article-title":"Emotion, cognition, and behavior","volume":"298","author":"Dolan","year":"2002","journal-title":"science"},{"key":"10.1016\/j.asoc.2026.114998_bib0045","first-page":"16","article-title":"Basic emotions","volume":"98","author":"Ekman","year":"1999","journal-title":"Handb. Cogn. Emot."},{"key":"10.1016\/j.asoc.2026.114998_bib0050","series-title":"Emotion, Psychopathology, and Psychotherapy","first-page":"3","article-title":"Emotions and psychotherapy: a psychoevolutionary perspective","author":"Plutchik","year":"1990"},{"key":"10.1016\/j.asoc.2026.114998_bib0055","series-title":"Proceedings of the 33rd ACM International Conference on Multimedia","first-page":"7064","article-title":"Eemo-bench: a benchmark for multi-modal large language models on image evoked emotion assessment","author":"Gao","year":"2025"},{"key":"10.1016\/j.asoc.2026.114998_bib0060","author":"Wu"},{"key":"10.1016\/j.asoc.2026.114998_bib0065","doi-asserted-by":"crossref","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","article-title":"Deep facial expression recognition: a survey","volume":"13","author":"Li","year":"2020","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.asoc.2026.114998_bib0070","doi-asserted-by":"crossref","first-page":"374","DOI":"10.1109\/TAFFC.2017.2714671","article-title":"Emotions recognition using EEG signals: a survey","volume":"10","author":"Alarcao","year":"2017","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.asoc.2026.114998_bib0075","doi-asserted-by":"crossref","first-page":"443","DOI":"10.1002\/hbm.23041","article-title":"Differences in neural activity when processing emotional arousal and valence in autism spectrum disorders","volume":"37","author":"Tseng","year":"2016","journal-title":"Hum. Brain Mapp."},{"key":"10.1016\/j.asoc.2026.114998_bib0080","doi-asserted-by":"crossref","first-page":"59","DOI":"10.1016\/j.neuropharm.2016.08.020","article-title":"Emotional arousal state influences the ability of amygdalar endocannabinoid signaling to modulate anxiety","volume":"111","author":"Morena","year":"2016","journal-title":"Neuropharmacology"},{"key":"10.1016\/j.asoc.2026.114998_bib0085","doi-asserted-by":"crossref","DOI":"10.2196\/13869","article-title":"Exposure therapy with personalized real-time arousal detection and feedback to alleviate social anxiety symptoms in an analogue adult sample: pilot proof-of-concept randomized controlled trial","volume":"6","author":"Lin","year":"2019","journal-title":"JMIR Ment. Health"},{"key":"10.1016\/j.asoc.2026.114998_bib0090","doi-asserted-by":"crossref","first-page":"e1","DOI":"10.1017\/S0140525X14000041","article-title":"Memory reconsolidation, emotional arousal, and the process of change in psychotherapy: new insights from brain science","volume":"38","author":"Lane","year":"2015","journal-title":"Behav. Brain Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0095","series-title":"2015 28th SIBGRAPI Conference on Graphics, Patterns and Images","first-page":"273","article-title":"A facial expression recognition system using convolutional networks","author":"Lopes","year":"2015"},{"key":"10.1016\/j.asoc.2026.114998_bib0100","doi-asserted-by":"crossref","first-page":"1993","DOI":"10.1109\/TMI.2014.2377694","article-title":"The multimodal brain tumor image segmentation benchmark (brats)","volume":"34","author":"Menze","year":"2014","journal-title":"IEEE Trans. Med. Imaging"},{"key":"10.1016\/j.asoc.2026.114998_bib0105","series-title":"2017 Seventh International Conference on Innovative Computing Technology (INTECH)","first-page":"46","article-title":"Real-time emotional state detection from facial expression on embedded devices","author":"Turabzadeh","year":"2017"},{"key":"10.1016\/j.asoc.2026.114998_bib0110","article-title":"Text-based emotion detection: advances, challenges, and opportunities","volume":"2","author":"Acheampong","year":"2020","journal-title":"Eng. Rep."},{"key":"10.1016\/j.asoc.2026.114998_bib0115","series-title":"2016 IEEE Annual India Conference (INDICON)","first-page":"1","article-title":"Emotion detection using perceptual based speech features","author":"Lalitha","year":"2016"},{"key":"10.1016\/j.asoc.2026.114998_bib0120","doi-asserted-by":"crossref","first-page":"593","DOI":"10.1016\/j.ins.2021.10.005","article-title":"A survey on facial emotion recognition techniques: a state-of-the-art literature review","volume":"582","author":"Canal","year":"2022","journal-title":"Inf. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0125","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1016\/j.entcs.2019.04.009","article-title":"Emotion recognition from physiological signal analysis: a review","volume":"343","author":"Egger","year":"2019","journal-title":"Electron. Notes Theor. Comput. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0130","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2021.102755","article-title":"EEG based emotion detection using fourth order spectral moment and deep learning","volume":"68","author":"Joshi","year":"2021","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.asoc.2026.114998_bib0135","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1007\/s12559-017-9517-x","article-title":"Attentional bias pattern recognition in spiking neural networks from spatio-temporal EEG data","volume":"10","author":"Gholami Doborjeh","year":"2018","journal-title":"Cogn. Comput."},{"key":"10.1016\/j.asoc.2026.114998_bib0140","series-title":"2022 9th International Conference on Computing for Sustainable Global Development (INDIACom)","first-page":"530","article-title":"Deep learning and machine learning based facial emotion detection using CNN","author":"Singh","year":"2022"},{"key":"10.1016\/j.asoc.2026.114998_bib0145","series-title":"2020 IEEE International Conference on Advent Trends in Multidisciplinary Research and Innovation (ICATMRI)","first-page":"1","article-title":"Emotion detection using deep facial features","author":"Kondaveeti","year":"2020"},{"key":"10.1016\/j.asoc.2026.114998_bib0150","doi-asserted-by":"crossref","first-page":"1527","DOI":"10.1016\/j.procs.2020.04.163","article-title":"Deep self-attention network for facial emotion recognition","volume":"171","author":"Gupta","year":"2020","journal-title":"Proc. Comput. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0155","doi-asserted-by":"crossref","first-page":"446","DOI":"10.1007\/s42452-020-2234-1","article-title":"Facial emotion recognition using convolutional neural networks (FERC)","volume":"2","author":"Mehendale","year":"2020","journal-title":"SN Appl. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0160","series-title":"2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS)","first-page":"317","article-title":"Facial emotion recognition using deep convolutional neural network","author":"Pranav","year":"2020"},{"key":"10.1016\/j.asoc.2026.114998_bib0165","doi-asserted-by":"crossref","first-page":"1036","DOI":"10.3390\/electronics10091036","article-title":"Facial emotion recognition using transfer learning in the deep CNN","volume":"10","author":"Akhand","year":"2021","journal-title":"Electronics"},{"key":"10.1016\/j.asoc.2026.114998_bib0170","article-title":"Development of a real-time emotion recognition system using facial expressions and EEG based on machine learning and deep neural network methods","volume":"20","author":"Hassouneh","year":"2020","journal-title":"Inf. Med. Unlocked"},{"key":"10.1016\/j.asoc.2026.114998_bib0175","doi-asserted-by":"crossref","first-page":"641","DOI":"10.1007\/s12530-023-09506-z","article-title":"Facial emotion recognition and music recommendation system using cnn-based deep learning techniques","volume":"15","author":"Bakariya","year":"2024","journal-title":"Evolving Systems"},{"key":"10.1016\/j.asoc.2026.114998_bib0180","series-title":"2020 IEEE 44th Annual Computers, Software, and Applications Conference (COMPSAC)","first-page":"408","article-title":"Finding emotion from multi-lingual voice data","author":"Hossain","year":"2020"},{"key":"10.1016\/j.asoc.2026.114998_bib0185","series-title":"ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"7379","article-title":"Comparison of glottal closure instants detection algorithms for emotional speech","author":"Kadiri","year":"2020"},{"key":"10.1016\/j.asoc.2026.114998_bib0190","series-title":"2021 IEEE 45th Annual Computers, Software, and Applications Conference (COMPSAC)","first-page":"338","article-title":"One source to detect them all: gender, age, and emotion detection from voice","author":"Zaman","year":"2021"},{"key":"10.1016\/j.asoc.2026.114998_bib0195","doi-asserted-by":"crossref","first-page":"4471","DOI":"10.1109\/TMM.2021.3118881","article-title":"I-gcn: incremental graph convolution network for conversation emotion detection","volume":"24","author":"Nie","year":"2021","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.asoc.2026.114998_bib0200","doi-asserted-by":"crossref","first-page":"20727","DOI":"10.1109\/ACCESS.2022.3149214","article-title":"Adaptive multimodal emotion detection architecture for social robots","volume":"10","author":"Heredia","year":"2022","journal-title":"IEEE Access"},{"key":"10.1016\/j.asoc.2026.114998_bib0205","doi-asserted-by":"crossref","first-page":"1072","DOI":"10.1080\/0144929X.2020.1741684","article-title":"Automatic voice emotion recognition of child-parent conversations in natural settings","volume":"40","author":"Law","year":"2021","journal-title":"Behav. Inf. Technol."},{"key":"10.1016\/j.asoc.2026.114998_bib0210","doi-asserted-by":"crossref","first-page":"47795","DOI":"10.1109\/ACCESS.2021.3068045","article-title":"A comprehensive review of speech emotion recognition systems","volume":"9","author":"Wani","year":"2021","journal-title":"IEEE Access"},{"key":"10.1016\/j.asoc.2026.114998_bib0215","doi-asserted-by":"crossref","first-page":"18416","DOI":"10.1109\/ACCESS.2024.3356357","article-title":"Challenges and opportunities of text-based emotion detection: a survey","volume":"12","author":"Al Maruf","year":"2024","journal-title":"IEEE Access"},{"key":"10.1016\/j.asoc.2026.114998_bib0220","series-title":"2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS)","first-page":"1339","article-title":"Facial emotion recognition using convolution neural network","author":"Modi","year":"2021"},{"key":"10.1016\/j.asoc.2026.114998_bib0225","first-page":"18661","article-title":"Supervised contrastive learning","volume":"33","author":"Khosla","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.asoc.2026.114998_bib0230","unstructured":"S.I. Shahid, Feature mapping algorithm, 2025."},{"key":"10.1016\/j.asoc.2026.114998_bib0235","series-title":"International Conference on Machine Learning","first-page":"28492","article-title":"Robust speech recognition via large-scale weak supervision","author":"Radford","year":"2023"},{"key":"10.1016\/j.asoc.2026.114998_bib0240","series-title":"2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018)","first-page":"59","article-title":"Openface 2.0: facial behavior analysis toolkit","author":"Baltrusaitis","year":"2018"},{"key":"10.1016\/j.asoc.2026.114998_bib0245","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0196391","article-title":"The ryerson audio-visual database of emotional speech and song (ravdess): a dynamic, multimodal set of facial and vocal expressions in north American English","volume":"13","author":"Livingstone","year":"2018","journal-title":"PLoS One"},{"key":"10.1016\/j.asoc.2026.114998_bib0250","doi-asserted-by":"crossref","first-page":"2251","DOI":"10.1080\/17470211003721642","article-title":"Perceptual cues in nonverbal vocal expressions of emotion","volume":"63","author":"Sauter","year":"2010","journal-title":"Q. J. Exp. Psychol."},{"key":"10.1016\/j.asoc.2026.114998_bib0255","doi-asserted-by":"crossref","first-page":"25","DOI":"10.1177\/1754073920930791","article-title":"The evolution of human vocal emotion","volume":"13","author":"Bryant","year":"2021","journal-title":"Emot. Rev."},{"key":"10.1016\/j.asoc.2026.114998_bib0260","doi-asserted-by":"crossref","first-page":"199","DOI":"10.1016\/j.neubiorev.2020.01.006","article-title":"The Neuroscience of sadness: a multidisciplinary synthesis and collaborative review","volume":"111","author":"Arias","year":"2020","journal-title":"Neurosci. Biobehav. Rev."},{"key":"10.1016\/j.asoc.2026.114998_bib0265","doi-asserted-by":"crossref","first-page":"146","DOI":"10.1007\/s12124-011-9177-4","article-title":"In the beginning was the familiar voice: personally familiar voices in the evolutionary and contemporary Biology of communication","volume":"46","author":"Sidtis","year":"2012","journal-title":"Integr. Psychol. Behav. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0270","doi-asserted-by":"crossref","first-page":"838","DOI":"10.1037\/a0017810","article-title":"The voice conveys specific emotions: evidence from vocal burst displays","volume":"9","author":"Simon-Thomas","year":"2009","journal-title":"Emotion"},{"key":"10.1016\/j.asoc.2026.114998_bib0275","article-title":"Empathy and compassion toward other species decrease with evolutionary divergence time","volume":"9","author":"Miralles","year":"2019","journal-title":"Scientific Rep."},{"key":"10.1016\/j.asoc.2026.114998_bib0280","series-title":"ICIDSSD","article-title":"Mel frequency cepstral coefficient: a review","author":"Ali","year":"2020"},{"key":"10.1016\/j.asoc.2026.114998_bib0285","doi-asserted-by":"crossref","first-page":"124","DOI":"10.1037\/h0030377","article-title":"Constants across cultures in the face and emotion","volume":"17","author":"Ekman","year":"1971","journal-title":"J. Pers. Soc. Psychol."},{"key":"10.1016\/j.asoc.2026.114998_bib0290","first-page":"433","article-title":"Vocal expression of emotion","author":"Scherer","year":"2003","journal-title":"Handb. Affect. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0295","doi-asserted-by":"crossref","DOI":"10.1016\/j.isci.2024.109175","article-title":"Deep learning reveals what facial expressions mean to people in different cultures","volume":"27","author":"Brooks","year":"2024","journal-title":"Iscience"},{"key":"10.1016\/j.asoc.2026.114998_bib0300","doi-asserted-by":"crossref","first-page":"1248","DOI":"10.1080\/13506285.2013.835367","article-title":"Culture and facial expressions of emotion","volume":"21","author":"Jack","year":"2013","journal-title":"Vis. Cogn."},{"key":"10.1016\/j.asoc.2026.114998_bib0305","doi-asserted-by":"crossref","first-page":"870","DOI":"10.1037\/0022-3514.66.5.870","article-title":"Varieties of disgust faces and the structure of disgust","volume":"66","author":"Rozin","year":"1994","journal-title":"J. Pers. Soc. Psychol."},{"key":"10.1016\/j.asoc.2026.114998_bib0310","doi-asserted-by":"crossref","first-page":"860","DOI":"10.1037\/a0022758","article-title":"Happy mouth and sad eyes: scanning emotional facial expressions","volume":"11","author":"Eisenbarth","year":"2011","journal-title":"Emotion"},{"key":"10.1016\/j.asoc.2026.114998_bib0315","doi-asserted-by":"crossref","first-page":"2165","DOI":"10.1109\/TCYB.2014.2366468","article-title":"Oblique decision tree ensemble via multisurface proximal support vector machine","volume":"45","author":"Zhang","year":"2015","journal-title":"IEEE Trans. Cybern."},{"key":"10.1016\/j.asoc.2026.114998_bib0320","doi-asserted-by":"crossref","first-page":"2120","DOI":"10.1109\/TKDE.2013.49","article-title":"Clustering-based ensembles as an alternative to stacking","volume":"26","author":"Jurek","year":"2014","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"10.1016\/j.asoc.2026.114998_bib0325","doi-asserted-by":"crossref","first-page":"108","DOI":"10.1109\/TAFFC.2019.2930695","article-title":"What an \u201cehm\u201d leaks about you: mapping fillers into personality traits with quantum evolutionary feature selection algorithms","volume":"13","author":"Tayarani","year":"2022","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.asoc.2026.114998_bib0330","doi-asserted-by":"crossref","first-page":"510","DOI":"10.3390\/bioengineering9100510","article-title":"Audio-visual stress classification using cascaded RNN-LSTM networks","volume":"9","author":"Gupta","year":"2022","journal-title":"Bioengineering"},{"key":"10.1016\/j.asoc.2026.114998_bib0335","series-title":"2022 15th International Conference on Human System Interaction (HSI)","first-page":"1","article-title":"Scser: supervised contrastive learning for speech emotion recognition using transformers","author":"Alaparthi","year":"2022"},{"key":"10.1016\/j.asoc.2026.114998_bib0340","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2023.104676","article-title":"Multimodal emotion recognition using cross modal audio-video fusion with attention and deep metric learning","volume":"133","author":"Mocanu","year":"2023","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.asoc.2026.114998_bib0345","unstructured":"J.P.R. Dhanith, S. Venkatraman, V. Sharma, S. Malarvannan, M. Narendra, Multimodal emotion recognition using audio-video transformer fusion with cross attention, arXiv preprint (2024)."},{"key":"10.1016\/j.asoc.2026.114998_bib0350","doi-asserted-by":"crossref","first-page":"377","DOI":"10.1109\/TAFFC.2014.2336244","article-title":"Crema-d: crowd-sourced emotional multimodal actors dataset","volume":"5","author":"Cao","year":"2014","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.asoc.2026.114998_bib0355","series-title":"2023 International Conference on Electronics, Information, and Communication (ICEIC)","first-page":"1","article-title":"Coordvit: a novel method of improve vision transformer-based speech emotion recognition using coordinate information concatenate","author":"Kim","year":"2023"},{"key":"10.1016\/j.asoc.2026.114998_bib0360","article-title":"A transformer-cnn hybrid model for cognitive behavioral therapy in psychological assessment and intervention for enhanced diagnostic accuracy and treatment efficiency","volume":"14","author":"Vuyyuru","year":"2023","journal-title":"Int. J. Adv. Comput. Sci. Appl."},{"key":"10.1016\/j.asoc.2026.114998_bib0365","series-title":"2022 IEEE 14th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP)","first-page":"1","article-title":"Audio-video fusion with double attention for multimodal emotion recognition","author":"Mocanu","year":"2022"},{"key":"10.1016\/j.asoc.2026.114998_bib0370","series-title":"2021 IEEE International Midwest Symposium on Circuits and Systems (MWSCAS)","first-page":"611","article-title":"Video-audio emotion recognition based on feature fusion deep learning method","author":"Song","year":"2021"},{"key":"10.1016\/j.asoc.2026.114998_bib0375","doi-asserted-by":"crossref","first-page":"79861","DOI":"10.1109\/ACCESS.2020.2990405","article-title":"Clustering-based speech emotion recognition by incorporating learned features and deep bilstm","volume":"8","author":"Sajjad","year":"2020","journal-title":"IEEE Access"},{"key":"10.1016\/j.asoc.2026.114998_bib0380","series-title":"2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII)","first-page":"552","article-title":"Multimodal and temporal perception of audio-visual cues for emotion recognition","author":"Ghaleb","year":"2019"},{"key":"10.1016\/j.asoc.2026.114998_bib0385","author":"Ibrahim"},{"key":"10.1016\/j.asoc.2026.114998_bib0390","author":"Ristea"},{"key":"10.1016\/j.asoc.2026.114998_bib0395","article-title":"Speech emotion recognition using a multi-time-scale approach to feature aggregation and an ensemble of SVM classifiers","volume":"49","author":"Stefanowska","year":"2024","journal-title":"Arch. Acoust."},{"key":"10.1016\/j.asoc.2026.114998_bib0400","doi-asserted-by":"crossref","first-page":"9981","DOI":"10.3390\/app14219981","article-title":"Speech emotion recognition using transfer learning: integration of advanced speaker embeddings and image recognition models","volume":"14","author":"Jakubec","year":"2024","journal-title":"Appl. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0405","first-page":"1","article-title":"Multimodal emotion recognition based on a fusion of audiovisual information with temporal dynamics","author":"Salas-C\u00e1ceres","year":"2024","journal-title":"Multimed. Tools Appl."},{"key":"10.1016\/j.asoc.2026.114998_bib0410","doi-asserted-by":"crossref","first-page":"291","DOI":"10.1007\/s11263-024-02186-5","article-title":"Learning rate curriculum","volume":"133","author":"Croitoru","year":"2025","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.asoc.2026.114998_bib0415","doi-asserted-by":"crossref","first-page":"389","DOI":"10.1098\/rstb.2010.0117","article-title":"Disgust as an adaptive system for disease avoidance behaviour","volume":"366","author":"Curtis","year":"2011","journal-title":"Philos. Trans. R. Soc. B Biol. Sci."},{"key":"10.1016\/j.asoc.2026.114998_bib0420","series-title":"International Handbook of Anger: Constituent and Concomitant Biological, Psychological, and Social Processes","first-page":"361","article-title":"Madmen: an evolutionary perspective on anger and men\u2019s violent responses to transgression","author":"Fessler","year":"2009"},{"key":"10.1016\/j.asoc.2026.114998_bib0425","series-title":"International Conference on Machine Learning","first-page":"3319","article-title":"Axiomatic attribution for deep networks","author":"Sundararajan","year":"2017"},{"key":"10.1016\/j.asoc.2026.114998_bib0430","doi-asserted-by":"crossref","first-page":"203","DOI":"10.1093\/oso\/9780195169157.003.0014","article-title":"Observer-based measurement of facial expression with the facial action coding system","volume":"1","author":"Cohn","year":"2007","journal-title":"The Handbook of Emotion Elicitation and Assessment"},{"key":"10.1016\/j.asoc.2026.114998_bib0435","first-page":"45","article-title":"Methods for measuring facial action","author":"Ekman","year":"1982","journal-title":"Handbook of Methods in Nonverbal Behavior Research"},{"key":"10.1016\/j.asoc.2026.114998_bib0440","first-page":"1","article-title":"Measuring facial action","volume":"525","author":"Cohn","year":"2005","journal-title":"The New Handbook of Methods in Nonverbal Behavior Research"},{"key":"10.1016\/j.asoc.2026.114998_bib0445","doi-asserted-by":"crossref","DOI":"10.3389\/frobt.2020.532279","article-title":"Emotion recognition for human-robot interaction: recent advances and future perspectives","volume":"7","author":"Spezialetti","year":"2020","journal-title":"Front. Robot. AI"},{"key":"10.1016\/j.asoc.2026.114998_bib0450","doi-asserted-by":"crossref","DOI":"10.1016\/j.arr.2022.101633","article-title":"Socially assistive robots for people with dementia: systematic review and meta-analysis of feasibility, acceptability and the effect on cognition, neuropsychiatric symptoms and quality of life","volume":"78","author":"Yu","year":"2022","journal-title":"Ageing Res. Rev."},{"key":"10.1016\/j.asoc.2026.114998_bib0455","article-title":"User perspectives on emotionally aligned social robots for older adults and persons living with dementia","volume":"9","author":"Dosso","year":"2022","journal-title":"J. Rehabil. Assist. Technol. Eng."},{"key":"10.1016\/j.asoc.2026.114998_bib0460","doi-asserted-by":"crossref","first-page":"2020","DOI":"10.1109\/TAFFC.2022.3143803","article-title":"Artificial emotional intelligence in socially assistive robots for older adults: a pilot study","volume":"14","author":"Abdollahi","year":"2022","journal-title":"IEEE Trans. Affect. Comput."}],"container-title":["Applied Soft Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1568494626004461?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1568494626004461?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T16:06:30Z","timestamp":1776873990000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1568494626004461"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,7]]},"references-count":92,"alternative-id":["S1568494626004461"],"URL":"https:\/\/doi.org\/10.1016\/j.asoc.2026.114998","relation":{},"ISSN":["1568-4946"],"issn-type":[{"value":"1568-4946","type":"print"}],"subject":[],"published":{"date-parts":[[2026,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A deep learning feature mapping algorithm for emotion detection via facial and audio signals","name":"articletitle","label":"Article Title"},{"value":"Applied Soft Computing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.asoc.2026.114998","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 The Author(s). Published by Elsevier B.V.","name":"copyright","label":"Copyright"}],"article-number":"114998"}}