{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,24]],"date-time":"2026-04-24T15:01:26Z","timestamp":1777042886966,"version":"3.51.4"},"reference-count":150,"publisher":"Springer Science and Business Media LLC","issue":"35","license":[{"start":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T00:00:00Z","timestamp":1726272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T00:00:00Z","timestamp":1726272000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100016833","name":"Yuncheng University","doi-asserted-by":"publisher","award":["YY-202312"],"award-info":[{"award-number":["YY-202312"]}],"id":[{"id":"10.13039\/100016833","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s00521-024-10371-3","type":"journal-article","created":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T09:02:25Z","timestamp":1726304545000},"page":"21923-21956","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["A review of multimodal-based emotion recognition techniques for cyberbullying detection in online social media platforms"],"prefix":"10.1007","volume":"36","author":[{"given":"Shuai","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9114-7945","authenticated-orcid":false,"given":"Abdul Samad","family":"Shibghatullah","sequence":"additional","affiliation":[]},{"given":"Thirupattur Javid","family":"Iqbal","sequence":"additional","affiliation":[]},{"given":"Kay Hooi","family":"Keoy","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,14]]},"reference":[{"key":"10371_CR1","doi-asserted-by":"crossref","unstructured":"Shakeel N, Dwivedi RK (2022) A survey on detection of cyberbullying in social media using machine learning techniques. In: Intelligent communication technologies and virtual mobile networks: proceedings of ICICV 2022, Springer, pp 323\u2013340.","DOI":"10.1007\/978-981-19-1844-5_25"},{"issue":"Suppl 1","key":"10371_CR2","doi-asserted-by":"crossref","first-page":"1375","DOI":"10.1007\/s10462-023-10553-w","volume":"56","author":"V Balakrisnan","year":"2023","unstructured":"Balakrisnan V, Kaity M (2023) Cyberbullying detection and machine learning: a systematic literature review.\u00a0Artif Intell Rev 56(Suppl 1):1375\u20131416","journal-title":"Artif Intell Rev"},{"key":"10371_CR3","unstructured":"Hinduja S, Patchin JW (2021) Cyberbullying: identification, prevention, and response. Cyberbullying research center. https:\/\/cyberbullying.org\/what-is-cyberbullying. Accessed 20 May 2023"},{"issue":"4","key":"10371_CR4","doi-asserted-by":"crossref","first-page":"1073","DOI":"10.1037\/a0035618","volume":"140","author":"RM Kowalski","year":"2014","unstructured":"Kowalski RM, Giumetti GW, Schroeder AN, Lattanner MR (2014) Bullying in the digital age: a critical review and meta-analysis of cyberbullying research among youth. Psychol Bull 140(4):1073\u20131137","journal-title":"Psychol Bull"},{"key":"10371_CR5","doi-asserted-by":"crossref","first-page":"333","DOI":"10.1016\/j.chb.2018.12.021","volume":"93","author":"H Rosa","year":"2019","unstructured":"Rosa H, Pereira N, Ribeiro R, Ferreira PC, Carvalho JP, Oliveira S, Coheur L, Paulino P, Veiga Simao AM, Trancoso I (2019) Automatic cyberbullying detection: A systematic review. Comput Hum Behav 93:333\u2013345","journal-title":"Comput Hum Behav"},{"key":"10371_CR6","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2021.115001","volume":"179","author":"A Bozyi\u011fit","year":"2021","unstructured":"Bozyi\u011fit A, Utku S, Nasibov E (2021) Cyberbullying detection: Utilizing social media features. Expert Syst Appl 179:115001","journal-title":"Expert Syst Appl"},{"issue":"1","key":"10371_CR7","first-page":"7","volume":"10","author":"J Bishop","year":"2014","unstructured":"Bishop J (2014) Representations of \u2018trolls\u2019 in mass media communication: a review of media-texts and moral panics relating to \u2018internet trolling.\u2019 Int J Web Based 10(1):7\u201324","journal-title":"Int J Web Based"},{"issue":"6","key":"10371_CR8","doi-asserted-by":"crossref","first-page":"1409","DOI":"10.1007\/s10207-022-00600-y","volume":"21","author":"R Kumar","year":"2022","unstructured":"Kumar R, Bhat A (2022) A study of machine learning-based models for detection, control, and mitigation of cyberbullying in online social media. Int J Inf Secur 21(6):1409\u20131431","journal-title":"Int J Inf Secur"},{"issue":"3","key":"10371_CR9","doi-asserted-by":"crossref","first-page":"592","DOI":"10.3390\/s20030592","volume":"20","author":"A Dzedzickis","year":"2020","unstructured":"Dzedzickis A, Kaklauskas A, Bucinskas V (2020) Human emotion recognition: Review of sensors and methods. Sensors 20(3):592","journal-title":"Sensors"},{"key":"10371_CR10","doi-asserted-by":"crossref","unstructured":"Mehrabian A (2017) Communication without words. In:\u00a0Communication Theory, Routledge, pp 193\u2013200","DOI":"10.4324\/9781315080918-15"},{"key":"10371_CR11","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2022.108580","volume":"244","author":"AI Middya","year":"2022","unstructured":"Middya AI, Nag B, Roy S (2022) Deep learning based multimodal emotion recognition using model-level fusion of audio\u2013visual modalities. Knowledge Based Syst 244:108580","journal-title":"Knowledge Based Syst"},{"issue":"02","key":"10371_CR12","first-page":"52","volume":"2","author":"SMSA Abdullah","year":"2021","unstructured":"Abdullah SMSA, Ameen SYA, Sadeeq MA, Zeebaree S (2021) Multimodal emotion recognition using deep learning. J Appl Sci Technol Trends 2(02):52\u201358","journal-title":"J Appl Sci Technol Trends"},{"issue":"7","key":"10371_CR13","first-page":"1479","volume":"16","author":"XM Zhao","year":"2022","unstructured":"Zhao XM, Yang YJ, Zhang SQ (2022) Survey of deep learning based multimodal emotion recognition. J Front Comput Sci Technol 16(7):1479\u20131503","journal-title":"J Front Comput Sci Technol"},{"issue":"1","key":"10371_CR14","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/T-AFFC.2010.1","volume":"1","author":"RA Calvo","year":"2010","unstructured":"Calvo RA, D\u2019Mello S (2010) Affect detection: an interdisciplinary review of models, methods, and their applications. IEEE Trans Affect Comput 1(1):18\u201337","journal-title":"IEEE Trans Affect Comput"},{"key":"10371_CR15","doi-asserted-by":"crossref","DOI":"10.1016\/j.jnca.2019.102447","volume":"149","author":"NJ Shoumy","year":"2020","unstructured":"Shoumy NJ, Ang LM, Seng KP, Rahaman DM, Zia T (2020) Multimodal big data affective analytics: a comprehensive survey using text, audio, visual and physiological signals. J Netw Comput Appl 149:102447","journal-title":"J Netw Comput Appl"},{"issue":"3","key":"10371_CR16","doi-asserted-by":"crossref","first-page":"685","DOI":"10.1007\/s12525-021-00475-2","volume":"31","author":"C Janiesch","year":"2021","unstructured":"Janiesch C, Zschech P, Heinrich K (2021) Machine learning and deep learning. Electron Mark 31(3):685\u2013695","journal-title":"Electron Mark"},{"key":"10371_CR17","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1016\/j.inffus.2018.10.009","volume":"51","author":"MM Hassan","year":"2019","unstructured":"Hassan MM, Alam MGR, Uddin MZ, Huda S, Almogren A, Fortino G (2019) Human emotion recognition using deep belief network architecture. Inf Fusion 51:10\u201318","journal-title":"Inf Fusion"},{"issue":"1","key":"10371_CR18","doi-asserted-by":"crossref","first-page":"54","DOI":"10.1080\/23307706.2022.2085198","volume":"10","author":"PR Prakash","year":"2023","unstructured":"Prakash PR, Anuradha D, Iqbal J, Galety MG, Singh R, Neelakandan S (2023) A novel convolutional neural network with gated recurrent unit for automated speech emotion recognition and classification. J Control Decis 10(1):54\u201363","journal-title":"J Control Decis"},{"issue":"3","key":"10371_CR19","doi-asserted-by":"crossref","first-page":"1753","DOI":"10.1007\/s11831-021-09647-x","volume":"29","author":"SP Yadav","year":"2022","unstructured":"Yadav SP, Zaidi S, Mishra A, Yadav V (2022) Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Arch Comput Method Eng 29(3):1753\u20131770","journal-title":"Arch Comput Method Eng"},{"key":"10371_CR20","doi-asserted-by":"publisher","DOI":"10.46827\/ejes.v10i7.4856","author":"S Dimitrios","year":"2023","unstructured":"Dimitrios S, Vrakas G, Papadimitropoulou P (2023) School Bullying in high school students, prevention and coping strategies. Eur J Educ Stud. https:\/\/doi.org\/10.46827\/ejes.v10i7.4856","journal-title":"Eur J Educ Stud"},{"key":"10371_CR21","unstructured":"Nirmal N, Sable P, Patil P, Kuchiwale S (2021) Automated detection of cyberbullying using machine learning.\u00a0Int Res J Eng Technol (IRJET), 2054\u20132061."},{"key":"10371_CR22","doi-asserted-by":"publisher","DOI":"10.22214\/ijraset.2021.38701","author":"M Patidar","year":"2021","unstructured":"Patidar M, Lathi M, Jain M, Dhakad M, Barge Y (2021) Cyber bullying detection for twitter using ML classification algorithms. Int J Res Appl Sci Eng Technol (IJRASET). https:\/\/doi.org\/10.22214\/ijraset.2021.38701","journal-title":"Int J Res Appl Sci Eng Technol (IJRASET)"},{"key":"10371_CR23","volume":"45","author":"GW Giumetti","year":"2022","unstructured":"Giumetti GW, Kowalski RM (2022) Cyberbullying via social media and well-being. Curr Opin Psychol 45:101314","journal-title":"Curr Opin Psychol"},{"issue":"1","key":"10371_CR24","first-page":"342","volume":"13","author":"V Malpe","year":"2020","unstructured":"Malpe V, Vaikole S (2020) A comprehensive study on cyberbullying detection using machine learning approach. Int J Futur Gener Commun Netw 13(1):342\u2013351","journal-title":"Int J Futur Gener Commun Netw"},{"key":"10371_CR25","doi-asserted-by":"crossref","unstructured":"Nurrahmi H, Nurjanah D (2018) Indonesian twitter cyberbullying detection using text classification and user credibility. In\u00a02018 international conference on information and communications technology (ICOIACT), IEEE, pp 543\u2013548.","DOI":"10.1109\/ICOIACT.2018.8350758"},{"key":"10371_CR26","unstructured":"Justin W. Patchin, Sameer Hinduja (2024) Summary of our cyberbullying research (2007\u20132023). Cyberbullying research center. https:\/\/cyberbullying.org\/summary-of-our-cyberbullying-research. Accessed 30 May 2024"},{"key":"10371_CR27","unstructured":"Ditch, Label (2023) All the latest cyberbullying statistics for 2023. BroadbandSearch. https:\/\/www.broadbandsearch.net\/blog\/cyber-bullying-statistics#post-navigation-0. Accessed 6 June 2023"},{"issue":"5","key":"10371_CR28","first-page":"1","volume":"36","author":"YZ Wu","year":"2022","unstructured":"Wu YZ, Li HR, Yao T, He XD (2022) A survey of multimodal information processing frontiers: application, fusion and pre-training. J Chin Inf Process 36(5):1\u201320","journal-title":"J Chin Inf Process"},{"key":"10371_CR29","doi-asserted-by":"crossref","first-page":"114","DOI":"10.1016\/j.cviu.2015.09.015","volume":"147","author":"X Huang","year":"2016","unstructured":"Huang X, Kortelainen J, Zhao G, Li X, Moilanen A, Sepp\u00e4nen T, Pietik\u00e4inen M (2016) Multi-modal emotion analysis from facial expressions and electroencephalogram. Comput Vis Image Underst 147:114\u2013124","journal-title":"Comput Vis Image Underst"},{"issue":"6","key":"10371_CR30","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3457607","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2021) A survey on bias and fairness in machine learning. ACM Comput Surv 54(6):1\u201335","journal-title":"ACM Comput Surv"},{"key":"10371_CR31","doi-asserted-by":"crossref","first-page":"769","DOI":"10.1016\/j.future.2020.09.030","volume":"115","author":"M Arafeh","year":"2021","unstructured":"Arafeh M, Ceravolo P, Mourad A, Damiani E, Bellini E (2021) Ontology based recommender system using social network data. Future Gener Comput Syst 115:769\u2013779","journal-title":"Future Gener Comput Syst"},{"key":"10371_CR32","doi-asserted-by":"crossref","unstructured":"Dai W, Cahyawijaya S, Liu Z, Fung P (2021) Multimodal end-to-end sparse model for emotion recognition.\u00a0arXiv preprint arXiv:2103.09666.","DOI":"10.18653\/v1\/2021.naacl-main.417"},{"issue":"12","key":"10371_CR33","first-page":"22","volume":"101","author":"T Gulzar","year":"2014","unstructured":"Gulzar T, Singh A, Sharma S (2014) Comparative analysis of LPCC, MFCC and BFCC for the recognition of Hindi words using artificial neural networks. Int J Comput Appl 101(12):22\u201327","journal-title":"Int J Comput Appl"},{"key":"10371_CR34","doi-asserted-by":"crossref","first-page":"1039261","DOI":"10.3389\/fcomp.2023.1039261","volume":"5","author":"S Kshirsagar","year":"2023","unstructured":"Kshirsagar S, Pendyala A, Falk TH (2023) Task-specific speech enhancement and data augmentation for improved multimodal emotion recognition under noisy conditions. Front Comput Sci 5:1039261","journal-title":"Front Comput Sci"},{"key":"10371_CR35","doi-asserted-by":"crossref","unstructured":"Xu H, Zhang H, Han K, Wang Y, Peng Y, Li X (2019) Learning alignment for multimodal emotion recognition from speech.\u00a0arXiv preprint arXiv:1909.05645.","DOI":"10.21437\/Interspeech.2019-3247"},{"key":"10371_CR36","doi-asserted-by":"crossref","unstructured":"Adikara PP, Adinugroho S, Insani S (2020) Detection of cyber harassment (cyberbullying) on Instagram using na\u00efve bayes classifier with bag of words and lexicon based features. In: Proceedings of the 5th international conference on sustainable information engineering and technology 2020, pp 64\u201368.","DOI":"10.1145\/3427423.3427436"},{"key":"10371_CR37","doi-asserted-by":"crossref","unstructured":"Setiawan Y, Gunawan D, Efendi R (2022) Feature extraction TF-IDF to perform cyberbullying text classification: a literature review and future research direction. In:\u00a02022 international conference on information technology systems and innovation (ICITSI), IEEE, pp 283\u2013288","DOI":"10.1109\/ICITSI56531.2022.9970942"},{"issue":"1","key":"10371_CR38","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1017\/S1351324916000334","volume":"23","author":"KW Church","year":"2017","unstructured":"Church KW (2017) Word2Vec. Nat Lang Eng 23(1):155\u2013162","journal-title":"Nat Lang Eng"},{"key":"10371_CR39","doi-asserted-by":"crossref","unstructured":"Al-Hashedi M, Soon LK, Goh HN (2019) Cyberbullying detection using deep learning and word embeddings: an empirical study. In: Proceedings of the 2019 2nd international conference on computational intelligence and intelligent systems, 2019, pp 17\u201321.","DOI":"10.1145\/3372422.3373592"},{"issue":"5","key":"10371_CR40","doi-asserted-by":"crossref","DOI":"10.1016\/j.ipm.2021.102664","volume":"58","author":"L Shang","year":"2021","unstructured":"Shang L, Zhang Y, Zha Y, Chen Y, Youn C, Wang D (2021) Aomd: An analogy-aware approach to offensive meme detection on social media. Inf Process Manage 58(5):102664","journal-title":"Inf Process Manage"},{"key":"10371_CR41","doi-asserted-by":"crossref","unstructured":"Maity K, Jha P, Saha S, Bhattacharyya P (2022) A multitask framework for sentiment, emotion and sarcasm aware cyberbullying detection from multi-modal code-mixed memes. In: Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pp. 1739\u20131749.","DOI":"10.1145\/3477495.3531925"},{"key":"10371_CR42","doi-asserted-by":"crossref","unstructured":"Eyben F, Weninger F, Gross F, Schuller B (2013) Recent developments in opensmile, the munich open-source multimedia feature extractor. In: Proceedings of the 21st ACM international conference on Multimedia, pp 835\u2013838","DOI":"10.1145\/2502081.2502224"},{"key":"10371_CR43","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s11042-023-14850-y","volume":"82","author":"V Dwivedy","year":"2023","unstructured":"Dwivedy V, Roy PK (2023) Deep feature fusion for hate speech detection: a transfer learning approach. Multimed Tools Appl 82:1\u201323","journal-title":"Multimed Tools Appl"},{"key":"10371_CR44","doi-asserted-by":"crossref","unstructured":"Jia Z, Lin Y, Wang J, Feng Z, Xie X, Chen C (2021) HetEmotionNet: two-stream heterogeneous graph recurrent neural network for multi-modal emotion recognition. In: Proceedings of the 29th ACM international conference on multimedia, pp 1047\u20131056","DOI":"10.1145\/3474085.3475583"},{"issue":"8","key":"10371_CR45","doi-asserted-by":"crossref","first-page":"3549","DOI":"10.1109\/TKDE.2020.3028705","volume":"34","author":"Q Guo","year":"2020","unstructured":"Guo Q, Zhuang F, Qin C, Zhu H, Xie X, Xiong H, He Q (2020) A survey on knowledge graph-based recommender systems. IEEE Trans Knowl Data Eng 34(8):3549\u20133568","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"10371_CR46","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1007\/s12193-016-0222-y","volume":"11","author":"C Torres-Valencia","year":"2017","unstructured":"Torres-Valencia C, \u00c1lvarez-L\u00f3pez M, Orozco-Guti\u00e9rrez \u00c1 (2017) SVM-based feature selection methods for emotion recognition from multimodal data. J Multimodal User Interfaces 11:9\u201323","journal-title":"J Multimodal User Interfaces"},{"issue":"3","key":"10371_CR47","doi-asserted-by":"crossref","first-page":"1903","DOI":"10.1007\/s12652-021-03407-2","volume":"14","author":"B Pan","year":"2023","unstructured":"Pan B, Hirota K, Jia Z et al (2023) Multimodal emotion recognition based on feature selection and extreme learning machine in video clips. J Ambient Intell Human Comput 14(3):1903\u20131917","journal-title":"J Ambient Intell Human Comput"},{"key":"10371_CR48","doi-asserted-by":"crossref","unstructured":"Sharupa NA, Rahman M, Alvi N, et al. (2020) Emotion detection of Twitter post using multinomial Naive Bayes. In: 2020 11th international conference on computing, communication and networking technologies (ICCCNT). IEEE, pp 1\u20136","DOI":"10.1109\/ICCCNT49239.2020.9225432"},{"key":"10371_CR49","doi-asserted-by":"crossref","first-page":"271","DOI":"10.1016\/j.neucom.2017.07.050","volume":"273","author":"ZT Liu","year":"2018","unstructured":"Liu ZT, Wu M, Cao WH et al (2018) Speech emotion recognition based on feature selection and extreme learning machine decision tree. Neurocomputing 273:271\u2013280","journal-title":"Neurocomputing"},{"key":"10371_CR50","doi-asserted-by":"crossref","first-page":"150","DOI":"10.1016\/j.ins.2019.09.005","volume":"509","author":"L Chen","year":"2020","unstructured":"Chen L, Su W, Feng Y et al (2020) Two-layer fuzzy multiple random forest for speech emotion recognition in human-robot interaction. Inf Sci 509:150\u2013163","journal-title":"Inf Sci"},{"key":"10371_CR51","doi-asserted-by":"crossref","unstructured":"He X, Huang J, Zeng Z (2021) Logistic regression based multi-task, multi-kernel learning for emotion recognition. In: 6th IEEE International conference on advanced robotics and mechatronics (ICARM), pp 572\u2013577","DOI":"10.1109\/ICARM52023.2021.9536130"},{"key":"10371_CR52","unstructured":"Kusal S, Patil S, Choudrie J, et al. (2023) A review on text-based emotion detection--techniques, applications, datasets, and future directions. arXiv preprint arXiv:2205.03235."},{"issue":"3\u20134","key":"10371_CR53","doi-asserted-by":"crossref","first-page":"197","DOI":"10.1561\/2000000039","volume":"7","author":"L Deng","year":"2014","unstructured":"Deng L, Yu D (2014) Deep learning: methods and applications. Found Trends\u00ae Signal Process 7(3\u20134):197\u2013387","journal-title":"Found Trends\u00ae Signal Process"},{"key":"10371_CR54","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, et al (2014) Generative adversarial nets. Adv Neural Inf Process Syst 27. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2014\/file\/5ca3e9b122f61f8f06494c97b1afccf3-Paper.pdf"},{"issue":"1","key":"10371_CR55","first-page":"261","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need.\u00a0Adv Neural Inf Process Syst\u00a030(1):261\u2013272","journal-title":"Adv Neural Inf Process Syst"},{"key":"10371_CR56","first-page":"1","volume":"1","author":"M Munikar","year":"2019","unstructured":"Munikar M, Shakya S (2019) Shrestha A (2019) Fine-grained sentiment classification using BERT. Artif Intell Transform Bus Soc (AITB) 1:1\u20135","journal-title":"Artif Intell Transform Bus Soc (AITB)"},{"key":"10371_CR57","unstructured":"Devlin J, Chang MW, Lee K, Toutanova, K (2018) BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018."},{"key":"10371_CR58","volume":"218","author":"MR Ahmed","year":"2023","unstructured":"Ahmed MR, Islam S, Islam AM, Shatabda S (2023) An ensemble 1D-CNN-LSTM-GRU model with data augmentation for speech emotion recognition. Expert Syst Appl 218:119633","journal-title":"Expert Syst Appl"},{"key":"10371_CR59","doi-asserted-by":"crossref","first-page":"39","DOI":"10.1016\/j.procs.2022.12.109","volume":"216","author":"R Febrian","year":"2023","unstructured":"Febrian R, Halim BM, Christina M, Ramdhan D, Chowanda A (2023) Facial expression recognition using bidirectional LSTM-CNN. Procedia Comput Sci 216:39\u201347","journal-title":"Procedia Comput Sci"},{"issue":"6","key":"10371_CR60","first-page":"1","volume":"6","author":"MG Huddar","year":"2021","unstructured":"Huddar MG, Sannakki SS (2021) Rajpurohit VS (2021) Attention-based multi-modal sentiment analysis and emotion detection in conversation using RNN. Int J Interact Multimed Artif Intell 6(6):1\u201310","journal-title":"Int J Interact Multimed Artif Intell"},{"key":"10371_CR61","doi-asserted-by":"crossref","first-page":"61672","DOI":"10.1109\/ACCESS.2020.2984368","volume":"8","author":"NH Ho","year":"2020","unstructured":"Ho NH, Yang HJ, Kim SH et al (2020) Multimodal approach of speech emotion recognition using multi-level multi-head fusion attention-based recurrent neural network. IEEE Access 8:61672\u201361686","journal-title":"IEEE Access"},{"key":"10371_CR62","doi-asserted-by":"crossref","unstructured":"Joshi A, Bhat A, Jain A, et al. (2020) COGMEN: Contextualized GNN based multimodal emotion recognition. In: Proceedings of the 2022 conference of the North American chapter of the association for computational linguistics: human language technologies 2022: 4148\u20134164.","DOI":"10.18653\/v1\/2022.naacl-main.306"},{"issue":"1","key":"10371_CR63","doi-asserted-by":"crossref","first-page":"527","DOI":"10.3390\/app12010527","volume":"12","author":"F Ma","year":"2022","unstructured":"Ma F, Li Y, Ni S et al (2022) Data augmentation for audio-visual emotion recognition with an efficient multimodal conditional GAN. Appl Sci 12(1):527","journal-title":"Appl Sci"},{"key":"10371_CR64","volume":"270","author":"K Mustaqeem","year":"2023","unstructured":"Mustaqeem K, El Saddik A, Alotaibi FS, Pham NT (2023) AAD-Net: Advanced end-to-end signal processing system for human emotion detection and recognition using attention-based deep echo state network. Knowl Based Syst 270:110525","journal-title":"Knowl Based Syst"},{"issue":"1","key":"10371_CR65","first-page":"539","volume":"135","author":"T Shen","year":"2023","unstructured":"Shen T, Xu H (2023) Facial expression recognition based on multi-channel attention residual network. Comp Model Eng Sci 135(1):539\u2013560","journal-title":"Comp Model Eng Sci"},{"key":"10371_CR66","doi-asserted-by":"crossref","unstructured":"Krishna DN, Patil A (2020) Multimodal emotion recognition using cross-modal attention and 1D convolutional neural networks. In: Interspeech, pp 4243\u20134247","DOI":"10.21437\/Interspeech.2020-1190"},{"key":"10371_CR67","volume":"202","author":"ZT Liu","year":"2023","unstructured":"Liu ZT, Han MT, Wu BH et al (2023) Speech emotion recognition based on convolutional neural network with attention-based bidirectional long short-term memory network and multi-task learning. Appl Acoust 202:109178","journal-title":"Appl Acoust"},{"issue":"5","key":"10371_CR68","first-page":"565","volume":"59","author":"JJ Liu","year":"2020","unstructured":"Liu JJ, Wu XF (2020) Real-time multimodal emotion recognition and emotion space labeling using LSTM networks. J Fudan Univ: Nat Sci 59(5):565\u2013574","journal-title":"J Fudan Univ: Nat Sci"},{"issue":"5","key":"10371_CR69","doi-asserted-by":"crossref","first-page":"1","DOI":"10.18637\/jss.v028.i05","volume":"28","author":"M Kuhn","year":"2008","unstructured":"Kuhn M (2008) Building predictive models in R using the caret package. J Stat software 28(5):1\u201326","journal-title":"J Stat software"},{"issue":"4","key":"10371_CR70","doi-asserted-by":"crossref","first-page":"427","DOI":"10.1016\/j.ipm.2009.03.002","volume":"45","author":"M Sokolova","year":"2009","unstructured":"Sokolova M, Lapalme G (2009) A systematic analysis of performance measures for classification tasks. Inf Process manage 45(4):427\u2013437","journal-title":"Inf Process manage"},{"key":"10371_CR71","unstructured":"Powers DM (2020) Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation.\u00a0arXiv preprint arXiv:2010.16061."},{"key":"10371_CR72","doi-asserted-by":"crossref","unstructured":"Emon MIH, Iqbal KN, Mehedi MHK, Mahbub MJA, Rasel AA (2022) Detection of Bangla hate comments and cyberbullying in social media using NLP and transformer models. In: ICACDS 2022, communications in computer and information science, 1613: 86-96. Springer, Cham","DOI":"10.1007\/978-3-031-12638-3_8"},{"issue":"3","key":"10371_CR73","doi-asserted-by":"crossref","first-page":"5307","DOI":"10.32604\/cmc.2023.031848","volume":"75","author":"KMO Nahar","year":"2023","unstructured":"Nahar KMO, Alauthman M, Yonbawi S, Almomani A (2023) Cyberbullying detection and recognition with type determination based on machine learning. Comput Mater Continua 75(3):5307\u20135319","journal-title":"Comput Mater Continua"},{"key":"10371_CR74","doi-asserted-by":"crossref","unstructured":"Davis J, Goadrich M (2006) The relationship between precision-recall and ROC curves. In: Proceedings of the 23rd international conference on Machine learning\u00a02006:233\u2013240.","DOI":"10.1145\/1143844.1143874"},{"key":"10371_CR75","doi-asserted-by":"crossref","first-page":"135","DOI":"10.1162\/tacl_a_00051","volume":"5","author":"P Bojanowski","year":"2017","unstructured":"Bojanowski P, Grave E, Joulin A (2017) Mikolov T (2017) Enriching word vectors with subword information. Transact Assoc Comput Linguist 5:135\u2013146","journal-title":"Transact Assoc Comput Linguist"},{"issue":"3","key":"10371_CR76","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1109\/MMUL.2012.26","volume":"19","author":"A Dhall","year":"2012","unstructured":"Dhall A, Goecke R, Lucey S, Gedeon T (2012) Collecting large, richly annotated facial-expression databases from movies. IEEE Multimedia 19(3):34","journal-title":"IEEE Multimedia"},{"issue":"3","key":"10371_CR77","doi-asserted-by":"crossref","first-page":"300","DOI":"10.1109\/TAFFC.2016.2553038","volume":"8","author":"S Zhalehpour","year":"2016","unstructured":"Zhalehpour S, Onder O, Akhtar Z, Erdem CE (2016) BAUM-1: A spontaneous audio-visual face database of affective and mental states. IEEE Trans Affective Comput 8(3):300\u2013313","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR78","doi-asserted-by":"crossref","first-page":"913","DOI":"10.1007\/s12652-016-0406-z","volume":"8","author":"Y Li","year":"2017","unstructured":"Li Y, Tao J, Chao L, Bao W, Liu Y (2017) CHEAVD: a Chinese natural emotional audio\u2013visual database. J Ambient Intell Human Comput 8:913\u2013924","journal-title":"J Ambient Intell Human Comput"},{"key":"10371_CR79","doi-asserted-by":"crossref","unstructured":"Yu W, Xu H, Meng F, Zhu Y, Ma Y, Wu J, Zou J, Yang K (2020) Ch-sims: A chinese multimodal sentiment analysis dataset with fine-grained annotation of modality. In:\u00a0Proceedings of the 58th annual meeting of the association for computational linguistics,\u00a0pp 3718\u20133727.","DOI":"10.18653\/v1\/2020.acl-main.343"},{"key":"10371_CR80","unstructured":"Zadeh A, Zellers R, Pincus E, Morency LP (2016) Mosi: multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos.\u00a0arXiv preprint arXiv:1606.06259."},{"key":"10371_CR81","unstructured":"Zadeh AB, Liang PP, Poria S, Cambria E, Morency LP (2018) Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In: Proceedings of the 56th annual meeting of the association for computational linguistics 1:2236\u20132246."},{"key":"10371_CR82","doi-asserted-by":"crossref","first-page":"8669","DOI":"10.1007\/s00521-020-05616-w","volume":"33","author":"J Chen","year":"2021","unstructured":"Chen J, Wang C, Wang K, Yin C, Zhao C, Xu T, Zhang X, Huang Z, Liu M, Yang T (2021) HEU Emotion: a large-scale database for multimodal emotion recognition in the wild. Neural Comput Appl 33:8669\u20138685","journal-title":"Neural Comput Appl"},{"key":"10371_CR83","doi-asserted-by":"crossref","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","volume":"42","author":"C Busso","year":"2008","unstructured":"Busso C, Bulut M, Lee CC et al (2008) IEMOCAP: Interactive emotional dyadic motion capture database. Lang Resour Eval 42:335\u2013359","journal-title":"Lang Resour Eval"},{"key":"10371_CR84","doi-asserted-by":"crossref","unstructured":"Poria S, Hazarika D, Majumder N, Naik G, Cambria E, Mihalcea R (2018) Meld: A multimodal multi-party dataset for emotion recognition in conversations.\u00a0arXiv preprint arXiv:1810.02508.","DOI":"10.18653\/v1\/P19-1050"},{"key":"10371_CR85","first-page":"501","volume":"11096","author":"O Perepelkina","year":"2018","unstructured":"Perepelkina O, Kazimirova E (2018) Konstantinova M (2018) RAMAS: Russian multimodal corpus of dyadic interaction for affective computing. SPECOM 11096:501\u2013510","journal-title":"SPECOM"},{"issue":"5","key":"10371_CR86","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0196391","volume":"13","author":"SR Livingstone","year":"2018","unstructured":"Livingstone SR, Russo FA (2018) The Ryerson audio-visual database of emotional speech and song (RAVDESS): a dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 13(5):e0196391","journal-title":"PLoS ONE"},{"key":"10371_CR87","volume-title":"Surrey audio-visual expressed emotion (savee) database","author":"P Jackson","year":"2014","unstructured":"Jackson P, Haq S (2014) Surrey audio-visual expressed emotion (savee) database. University of Surrey, Guildford"},{"issue":"3","key":"10371_CR88","doi-asserted-by":"crossref","first-page":"532","DOI":"10.1109\/TAFFC.2018.2817622","volume":"11","author":"T Song","year":"2018","unstructured":"Song T, Zheng W, Song P, Cui Z (2018) EEG emotion recognition using dynamical graph convolutional neural networks. IEEE Trans Affective Comput 11(3):532\u2013541","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR89","doi-asserted-by":"crossref","first-page":"117","DOI":"10.1007\/978-3-642-42051-1_16","volume":"8228","author":"IJ Goodfellow","year":"2013","unstructured":"Goodfellow IJ et al (2013) Challenges in representation learning: a report on three machine learning contests. ICONIP 2013. Lect Notes Comput Sci 8228:117\u2013124","journal-title":"Lect Notes Comput Sci"},{"key":"10371_CR90","doi-asserted-by":"crossref","unstructured":"Lucey P, Cohn JF, Kanade T, Saragih J, Ambadar Z, Matthews I (2010) The extended cohn-kanade dataset (ck+): a complete dataset for action unit and emotion-specified expression. In:\u00a02010 IEEE computer society conference on computer vision and pattern recognition-workshops, pp\u00a094\u2013101.","DOI":"10.1109\/CVPRW.2010.5543262"},{"issue":"1","key":"10371_CR91","doi-asserted-by":"crossref","first-page":"42","DOI":"10.1109\/T-AFFC.2011.25","volume":"3","author":"M Soleymani","year":"2011","unstructured":"Soleymani M, Lichtenauer J, Pun T, Pantic M (2011) A multimodal database for affect recognition and implicit tagging. IEEE Trans Affective Comput 3(1):42\u201355","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR92","doi-asserted-by":"crossref","unstructured":"Fabian Benitez-Quiroz C, Srinivasan R, Martinez AM (2016) Emotionet: an accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5562\u20135570.","DOI":"10.1109\/CVPR.2016.600"},{"key":"10371_CR93","unstructured":"P\u00e9rez-Rosas V, Mihalcea R, Morency LP (2013) Utterance-level multimodal sentiment analysis. In\u00a0Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics 1:973\u2013982."},{"issue":"1","key":"10371_CR94","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1109\/T-AFFC.2011.20","volume":"3","author":"G McKeown","year":"2011","unstructured":"McKeown G, Valstar M, Cowie R, Pantic M, Schroder M (2011) The semaine database: annotated multimodal records of emotionally colored conversations between a person and a limited agent. IEEE Trans Affective Comput 3(1):5\u201317","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR95","doi-asserted-by":"crossref","unstructured":"Ringeval F, Sonderegger A, Sauer J, Lalanne D (2013) Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions. In\u00a02013 10th IEEE international conference and workshops on automatic face and gesture recognition (FG)\u00a02013: 1\u20138.","DOI":"10.1109\/FG.2013.6553805"},{"issue":"1","key":"10371_CR96","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","volume":"3","author":"S Koelstra","year":"2011","unstructured":"Koelstra S, Muhl C, Soleymani M et al (2011) Deap: a database for emotion analysis; using physiological signals. IEEE Trans Affective Comput 3(1):18\u201331","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR97","doi-asserted-by":"crossref","unstructured":"Xu N, Mao W, Chen G (2019) Multi-interactive memory network for aspect based multimodal sentiment analysis. In: Proceedings of the AAAI conference on artificial intelligence\u00a033(01):371\u2013378","DOI":"10.1609\/aaai.v33i01.3301371"},{"key":"10371_CR98","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1016\/j.specom.2022.11.005","volume":"146","author":"P Singh","year":"2023","unstructured":"Singh P, Sahidullah M, Saha G (2023) Modulation spectral features for speech emotion recognition using deep neural networks. Speech Commun 146:53\u201369","journal-title":"Speech Commun"},{"key":"10371_CR99","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.neucom.2023.01.002","volume":"528","author":"J De Lope","year":"2023","unstructured":"De Lope J, Gra\u00f1a M (2023) An ongoing review of speech emotion recognition. Neurocomputing 528:1\u201311","journal-title":"Neurocomputing"},{"issue":"6","key":"10371_CR100","doi-asserted-by":"crossref","first-page":"490","DOI":"10.1109\/TMM.2010.2051872","volume":"12","author":"I Luengo","year":"2010","unstructured":"Luengo I, Navas E, Hern\u00e1ez I (2010) Feature analysis and evaluation for automatic emotion identification in speech. IEEE Trans Multimed 12(6):490\u2013501","journal-title":"IEEE Trans Multimed"},{"issue":"8","key":"10371_CR101","doi-asserted-by":"crossref","first-page":"4750","DOI":"10.3390\/app13084750","volume":"13","author":"AS Alluhaidan","year":"2023","unstructured":"Alluhaidan AS, Saidani O, Jahangir R, Nauman MA, Neffati OS (2023) Speech emotion recognition through hybrid features and convolutional neural network. Appl Sci 13(8):4750","journal-title":"Appl Sci"},{"key":"10371_CR102","doi-asserted-by":"crossref","unstructured":"Ottl S, Amiriparian S, Gerczuk M, Karas V, Schuller B (2020) Group-level speech emotion recognition utilising deep spectrum features. In: Proceedings of the 2020 international conference on multimodal interaction, pp 821\u2013826.","DOI":"10.1145\/3382507.3417964"},{"issue":"1","key":"10371_CR103","doi-asserted-by":"crossref","first-page":"84","DOI":"10.20965\/jaciii.2023.p0084","volume":"27","author":"LP Hung","year":"2023","unstructured":"Hung LP, Alias S (2023) Beyond sentiment analysis: A review of recent trends in text-based sentiment analysis and emotion detection. J Adv Comput Intell Intell Inf 27(1):84\u201395","journal-title":"J Adv Comput Intell Intell Inf"},{"issue":"3","key":"10371_CR104","doi-asserted-by":"publisher","first-page":"415","DOI":"10.14569\/IJACSA.2023.0140347","volume":"14","author":"M Errami","year":"2023","unstructured":"Errami M, Ouassil MA, Rachidi R, Cherradi B, Hamida S, Raihani A (2023) Sentiment analysis on moroccan dialect based on ML and social media content detection. Int J Adv Comput Sci Appl 14(3):415\u2013425. https:\/\/doi.org\/10.14569\/IJACSA.2023.0140347","journal-title":"Int J Adv Comput Sci Appl"},{"key":"10371_CR105","doi-asserted-by":"crossref","unstructured":"Pennington J, Socher R, Manning CD (2014) Glove: global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1532\u20131543.","DOI":"10.3115\/v1\/D14-1162"},{"key":"10371_CR106","doi-asserted-by":"crossref","unstructured":"Peters ME, Neumann M, Lyyer M, Gardner M, Clark C, Lee K, and Zettlemoyer L (2018) Deep contextualized word representations. In: Proceedings of the 2018 conference of the North American chapter of the association for computational linguistics: human language technologies 2018: 2227\u20132237.","DOI":"10.18653\/v1\/N18-1202"},{"key":"10371_CR107","unstructured":"Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training. Preprint, 1\u201312. https:\/\/cdn.openai.com\/research-covers\/language-unsupervised\/language_understanding_paper.pdf"},{"key":"10371_CR108","doi-asserted-by":"publisher","first-page":"3497","DOI":"10.1109\/ICASSP40776.2020.9054438","volume":"2020","author":"YA Chung","year":"2020","unstructured":"Chung YA, Glass J (2020) Generative pre-training for speech with autoregressive predictive coding. In ICASSP 2020:3497\u20133501. https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9054438","journal-title":"In ICASSP"},{"key":"10371_CR109","doi-asserted-by":"crossref","unstructured":"Dai Z, Yang Z, Yang Y, Carbonell J, Le QV, Salakhutdinov R (2019) Transformer-xl: Attentive language models beyond a fixed-length context. arXiv preprint arXiv:1901.02860.","DOI":"10.18653\/v1\/P19-1285"},{"key":"10371_CR110","unstructured":"Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR, Le QV (2019) Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32."},{"key":"10371_CR111","unstructured":"Liu Y, Ott M, Goyal N, et al. (2019) Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692."},{"issue":"8","key":"10371_CR112","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. OpenAI blog 1(8):9","journal-title":"OpenAI blog"},{"key":"10371_CR113","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"key":"10371_CR114","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang L, Wu J, Jiang X et al (2022) Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst 35:27730\u201327744","journal-title":"Adv Neural Inf Process Syst"},{"key":"10371_CR115","unstructured":"Achiam J, Adler S, Agarwal S, et al (2023) Gpt-4 technical report. arXiv preprint arXiv:2303.08774"},{"key":"10371_CR116","doi-asserted-by":"crossref","unstructured":"Tang D, Wei F, Yang N, Zhou M, Liu T, Qin B (2014) Learning sentiment-specific word embedding for twitter sentiment classification. In: Proceedings of the 52nd annual meeting of the association for computational linguistics (Volume 1: Long Papers) 2014: 1555\u20131565","DOI":"10.3115\/v1\/P14-1146"},{"key":"10371_CR117","doi-asserted-by":"crossref","unstructured":"Felbo B, Mislove A, S\u00f8gaard A, Rahwan I, Lehmann S (2017) Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm. arXiv preprint arXiv:1708.00524","DOI":"10.18653\/v1\/D17-1169"},{"key":"10371_CR118","first-page":"2824","volume":"80","author":"P Naga","year":"2023","unstructured":"Naga P, Marri SD, Borreo R (2023) Facial emotion recognition methods, datasets and technologies: a literature survey. Mater Today: Proc 80:2824\u20132828","journal-title":"Mater Today: Proc"},{"issue":"2","key":"10371_CR119","doi-asserted-by":"crossref","first-page":"399","DOI":"10.1007\/s11760-022-02246-8","volume":"17","author":"AH Sham","year":"2023","unstructured":"Sham AH, Aktas K, Rizhinashvili D, Kuklianov D, Alisinanoglu F, Ofodile I, Anbarjafari G (2023) Ethical AI in facial expression analysis: racial bias. Signal Image Video P 17(2):399\u2013406","journal-title":"Signal Image Video P"},{"key":"10371_CR120","first-page":"1","volume":"2023","author":"X Liu","year":"2023","unstructured":"Liu X, Xu Z, Huang K (2023) Multimodal emotion recognition based on cascaded multichannel and hierarchical fusion. Comput Intell Neurosci 2023:1\u201318","journal-title":"Comput Intell Neurosci"},{"issue":"1","key":"10371_CR121","doi-asserted-by":"crossref","first-page":"1333","DOI":"10.1007\/s11042-022-13210-6","volume":"82","author":"S Annadurai","year":"2023","unstructured":"Annadurai S, Arock M, Vadivel A (2023) Real and fake emotion detection using enhanced boosted support vector machine algorithm. Multimed Tools Appl 82(1):1333\u20131353","journal-title":"Multimed Tools Appl"},{"issue":"4","key":"10371_CR122","first-page":"1777","volume":"15","author":"S Vignesh","year":"2023","unstructured":"Vignesh S, Savithadevi M, Sridevi M, Sridhar R (2023) A novel facial emotion recognition model using segmentation VGG-19 architecture. Int J Inf Technol 15(4):1777\u20131787","journal-title":"Int J Inf Technol"},{"issue":"02","key":"10371_CR123","first-page":"52","volume":"2","author":"SMSA Abdullah","year":"2021","unstructured":"Abdullah SMSA, Ameen SYA, Sadeeq MA, Zeebaree S (2021) Multimodal emotion recognition using deep learning. Int J Appl Sci Technol Trends 2(02):52\u201358","journal-title":"Int J Appl Sci Technol Trends"},{"key":"10371_CR124","volume":"149","author":"NJ Shoumy","year":"2020","unstructured":"Shoumy NJ, Ang LM, Seng KP, Rahaman DM, Zia T (2020) Multimodal big data affective analytics: a comprehensive survey using text, audio, visual and physiological signals. Int J Network Comput Appl 149:102447","journal-title":"Int J Network Comput Appl"},{"key":"10371_CR125","doi-asserted-by":"publisher","unstructured":"Lv F et al (2021) Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition 2021:2554\u20132562. https:\/\/doi.org\/10.1109\/CVPR46437.2021.00258","DOI":"10.1109\/CVPR46437.2021.00258"},{"issue":"3","key":"10371_CR126","first-page":"1","volume":"47","author":"D\u2019mello SK, Kory J,","year":"2015","unstructured":"D\u2019mello SK, Kory J, (2015) A review and meta-analysis of multimodal affect detection systems. ACM Comput Surv (CSUR) 47(3):1\u201336","journal-title":"ACM Comput Surv (CSUR)"},{"key":"10371_CR127","unstructured":"Kumar R, Reganti AN, Bhatia A, Maheshwari T (2018) Aggression-annotated corpus of hindi-english code-mixed data. arXiv preprint arxiv:1803.09402"},{"key":"10371_CR128","first-page":"3507","volume":"2020","author":"J Huang","year":"2020","unstructured":"Huang J, Tao J, Liu B, Lian Z, Niu M (2020) Multimodal transformer fusion for continuous emotion recognition. In ICASSP 2020:3507\u20133511","journal-title":"In ICASSP"},{"key":"10371_CR129","doi-asserted-by":"crossref","first-page":"4489","DOI":"10.1007\/s12652-023-04567-z","volume":"14","author":"A Ghosh","year":"2023","unstructured":"Ghosh A, Dhara BC, Pero C et al (2023) A multimodal sentiment analysis system for recognizing person aggressiveness in pain based on textual and visual information. J Ambient Intell Human Comput 14:4489\u20134501","journal-title":"J Ambient Intell Human Comput"},{"key":"10371_CR130","doi-asserted-by":"crossref","first-page":"26989","DOI":"10.1007\/s11042-020-09631-w","volume":"81","author":"S Paul","year":"2022","unstructured":"Paul S, Saha S, Hasanuzzaman M (2022) Identification of cyberbullying: a deep learning based multimodal approach. Multimed Tools Appl 81:26989\u201327008","journal-title":"Multimed Tools Appl"},{"key":"10371_CR131","doi-asserted-by":"crossref","first-page":"985","DOI":"10.1109\/TASLP.2021.3049898","volume":"29","author":"Z Lian","year":"2021","unstructured":"Lian Z, Liu B, Tao J (2021) CTNet: Conversat transformer network for emotion recognition. IEEE\/ACM Trans Audio Speech Lang Process 29:985\u20131000","journal-title":"IEEE\/ACM Trans Audio Speech Lang Process"},{"key":"10371_CR132","doi-asserted-by":"crossref","first-page":"53907","DOI":"10.1109\/ACCESS.2023.3280556","volume":"11","author":"M Al-Hashedi","year":"2023","unstructured":"Al-Hashedi M, Soon LK, Goh HN, Lim AHL, Siew EG (2023) Cyberbullying Detection Based on Emotion. In IEEE Access 11:53907\u201353918","journal-title":"In IEEE Access"},{"key":"10371_CR133","doi-asserted-by":"crossref","first-page":"101921","DOI":"10.1016\/j.inffus.2023.101921","volume":"100","author":"T Yue","year":"2023","unstructured":"Yue T, Mao R, Wang H, Hu Z, Cambria E (2023) KnowleNet: knowledge fusion network for multimodal sarcasm detection. Inf Fusion 100:101921","journal-title":"Inf Fusion"},{"key":"10371_CR134","unstructured":"P\u00e9rez-Rosas V, Mihalcea R, Morency LP (2013) Utterance-level multi modal sentiment analysis. In\u00a0Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics 1:973\u2013982."},{"key":"10371_CR135","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.118523","volume":"211","author":"N Jaafar","year":"2023","unstructured":"Jaafar N, Lachiri Z (2023) Multimodal fusion methods with deep neural networks and meta-information for aggression detection in surveillance. Expert Syst Appl 211:118523","journal-title":"Expert Syst Appl"},{"issue":"1","key":"10371_CR136","doi-asserted-by":"crossref","first-page":"108","DOI":"10.1109\/TAFFC.2020.3038167","volume":"14","author":"S Poria","year":"2023","unstructured":"Poria S, Hazarika D, Majumder N, Mihalcea R (2023) Beneath the tip of the iceberg: current challenges and new directions in sentiment analysis research. IEEE Trans Affective Comput 14(1):108\u2013132","journal-title":"IEEE Trans Affective Comput"},{"issue":"3","key":"10371_CR137","doi-asserted-by":"crossref","first-page":"478","DOI":"10.1109\/JSTSP.2020.2987728","volume":"14","author":"C Zhang","year":"2020","unstructured":"Zhang C, Yang Z, He X, Deng L (2020) Multimodal intelligence: representation learning, information fusion, and applications. IEEE J Sel Top Signal Process 14(3):478\u2013493","journal-title":"IEEE J Sel Top Signal Process"},{"key":"10371_CR138","doi-asserted-by":"crossref","first-page":"1046","DOI":"10.1109\/LSP.2021.3078698","volume":"28","author":"M Ren","year":"2021","unstructured":"Ren M, Huang X, Shi X, Nie W (2021) Interactive multimodal attention network for emotion recognition in conversation. IEEE Signal Process Lett 28:1046\u20131050","journal-title":"IEEE Signal Process Lett"},{"key":"10371_CR139","doi-asserted-by":"crossref","first-page":"50","DOI":"10.1016\/j.neucom.2015.01.095","volume":"174","author":"S Poria","year":"2016","unstructured":"Poria S, Cambria E, Howard N, Huang GB, Hussain A (2016) Fusing audio, visual and textual clues for sentiment analysis from multimodal content. Neurocomputing 174:50\u201359","journal-title":"Neurocomputing"},{"issue":"1","key":"10371_CR140","doi-asserted-by":"crossref","first-page":"320","DOI":"10.1109\/TAFFC.2020.3000510","volume":"13","author":"S Mai","year":"2020","unstructured":"Mai S, Hu H, Xu J, Xing S (2020) Multi-fusion residual memory network for multimodal human sentiment comprehension. IEEE Trans Affective Comput 13(1):320\u2013334","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR141","doi-asserted-by":"crossref","unstructured":"Khare A, Parthasarathy S, Sundaram S (2021) Self-supervised learning with cross-modal transformers for emotion recognition. In:\u00a02021 IEEE spoken language technology workshop (SLT), pp 381\u2013388.","DOI":"10.1109\/SLT48900.2021.9383618"},{"issue":"1","key":"10371_CR142","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1109\/MSP.2017.2765202","volume":"35","author":"A Creswell","year":"2018","unstructured":"Creswell A, White T, Dumoulin V, Arulkumaran K, Sengupta B, Bharath AA (2018) Generative adversarial networks: an overview. IEEE Signal Process Mag 35(1):53\u201365","journal-title":"IEEE Signal Process Mag"},{"key":"10371_CR143","doi-asserted-by":"crossref","unstructured":"Mai S, Hu H, Xing S (2020) Modality to modality translation: An adversarial representation learning and graph fusion network for multimodal fusion. In: Proceedings of the AAAI conference on artificial intelligence\u00a034(01), pp 164\u2013172","DOI":"10.1609\/aaai.v34i01.5347"},{"key":"10371_CR144","unstructured":"Li H, Kadav A, Durdanovic I, Samet H, Graf HP (2016) Pruning filters for efficient convnets.\u00a0arXiv preprint arXiv:1608.08710."},{"key":"10371_CR145","doi-asserted-by":"publisher","unstructured":"He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In:\u00a0Proceedings of the IEEE international conference on computer vision, pp\u00a01389\u20131397. https:\/\/doi.org\/10.1109\/ICCV.2017.155","DOI":"10.1109\/ICCV.2017.155"},{"issue":"2","key":"10371_CR146","doi-asserted-by":"crossref","first-page":"894","DOI":"10.1109\/TAFFC.2020.2973984","volume":"13","author":"HJ Escalante","year":"2020","unstructured":"Escalante HJ, Kaya H, Salah AA et al (2020) Modeling, recognizing, and explaining apparent personality from videos. IEEE Trans Affective Comput 13(2):894\u2013911","journal-title":"IEEE Trans Affective Comput"},{"key":"10371_CR147","doi-asserted-by":"crossref","first-page":"185","DOI":"10.1016\/j.neunet.2020.07.010","volume":"130","author":"P Angelov","year":"2020","unstructured":"Angelov P, Soares E (2020) Towards explainable deep neural networks (xDNN). Neural Netw 130:185\u2013194","journal-title":"Neural Netw"},{"key":"10371_CR148","doi-asserted-by":"crossref","unstructured":"Yang C J, Fahier N, Li WC, Fang WC (2020) A convolution neural network based emotion recognition system using multimodal physiological signals. In:\u00a02020 IEEE International conference on consumer electronics-Taiwan (ICCE-Taiwan), pp 1\u20132.","DOI":"10.1109\/ICCE-Taiwan49838.2020.9258341"},{"issue":"2","key":"10371_CR149","doi-asserted-by":"crossref","first-page":"505","DOI":"10.1109\/TAFFC.2018.2874986","volume":"12","author":"F Noroozi","year":"2018","unstructured":"Noroozi F, Corneanu CA, Kami\u0144ska D, Sapi\u0144ski T, Escalera S, Anbarjafari G (2018) Survey on emotional body gesture recognition. IEEE Trans Affective Comput 12(2):505\u2013523","journal-title":"IEEE Trans Affective Comput"},{"issue":"5","key":"10371_CR150","doi-asserted-by":"crossref","first-page":"829","DOI":"10.1162\/neco_a_01273","volume":"32","author":"J Gao","year":"2020","unstructured":"Gao J, Li P, Chen Z, Zhang J (2020) A survey on deep learning for multimodal data fusion. Neural Comput 32(5):829\u2013864","journal-title":"Neural Comput"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10371-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-024-10371-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10371-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,25]],"date-time":"2024-11-25T12:03:14Z","timestamp":1732536194000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-024-10371-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,14]]},"references-count":150,"journal-issue":{"issue":"35","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["10371"],"URL":"https:\/\/doi.org\/10.1007\/s00521-024-10371-3","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9,14]]},"assertion":[{"value":"10 October 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 August 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 September 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All the authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}