{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T14:42:53Z","timestamp":1774968173083,"version":"3.50.1"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,6,7]],"date-time":"2023-06-07T00:00:00Z","timestamp":1686096000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,6,7]],"date-time":"2023-06-07T00:00:00Z","timestamp":1686096000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program","doi-asserted-by":"crossref","award":["2021YFB1600205"],"award-info":[{"award-number":["2021YFB1600205"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program","doi-asserted-by":"crossref","award":["2018YFB1600202"],"award-info":[{"award-number":["2018YFB1600202"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62172366"],"award-info":[{"award-number":["62172366"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52178407"],"award-info":[{"award-number":["52178407"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2023,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Driver emotion classification is an important topic that can raise awareness of driving habits because many drivers are overconfident and unaware of their bad driving habits. Drivers will acquire insight into their poor driving behaviors and be better able to avoid future accidents if their behavior is automatically identified. In this paper, we use different models such as convolutional neural networks, recurrent neural networks, and multi-layer perceptron classification models to construct an ensemble convolutional neural network-based enhanced driver facial expression recognition model. First, the faces of the drivers are discovered using the faster region-based convolutional neural network (R-CNN) model, which can recognize faces in real-time and offline video reliably and effectively. The feature-fusing technique is utilized to integrate the features extracted from three CNN models, and the fused features are then used to train the suggested ensemble classification model. To increase the accuracy and efficiency of face detection, a new convolutional neural network block (InceptionV3) replaces the improved Faster R-CNN feature-learning block. To evaluate the proposed face detection and driver facial expression recognition (DFER) datasets, we achieved an accuracy of 98.01%, 99.53%, 99.27%, 96.81%, and 99.90% on the JAFFE, CK+, FER-2013, AffectNet, and custom-developed datasets, respectively. The custom-developed dataset has been recorded as the best among all under the simulation environment.<\/jats:p>","DOI":"10.1007\/s40747-023-01100-9","type":"journal-article","created":{"date-parts":[[2023,6,7]],"date-time":"2023-06-07T04:34:31Z","timestamp":1686112471000},"page":"6927-6952","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":32,"title":["A novel driver emotion recognition system based on deep ensemble classification"],"prefix":"10.1007","volume":"9","author":[{"given":"Khalid","family":"Zaman","sequence":"first","affiliation":[]},{"given":"Sun","family":"Zhaoyun","sequence":"additional","affiliation":[]},{"given":"Babar","family":"Shah","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4761-0346","authenticated-orcid":false,"given":"Tariq","family":"Hussain","sequence":"additional","affiliation":[]},{"given":"Sayyed Mudassar","family":"Shah","sequence":"additional","affiliation":[]},{"given":"Farman","family":"Ali","sequence":"additional","affiliation":[]},{"given":"Umer Sadiq","family":"Khan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,7]]},"reference":[{"key":"1100_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.psychres.2022.115029","volume":"320","author":"J Pena-Garijo","year":"2023","unstructured":"Pena-Garijo J, Lacruz M, Masanet MJ, Palop-Grau A, Plaza R, Hernandez-Merino A, Valllina O (2023) Specific facial emotion recognition deficits across the course of psychosis: a comparison of individuals with low-risk, high-risk, first-episode psychosis and multi-episode schizophrenia-spectrum disorders. Psychiatry Res 320:115029","journal-title":"Psychiatry Res"},{"key":"1100_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.scitotenv.2019.135484","volume":"710","author":"Q Meng","year":"2020","unstructured":"Meng Q, Hu X, Kang J, Wu Y (2020) On the effectiveness of facial expression recognition for evaluation of urban sound perception. Sci Total Environ 710:135484","journal-title":"Sci Total Environ"},{"key":"1100_CR3","doi-asserted-by":"crossref","unstructured":"Hussain T, Yang B, Rahman HU, Iqbal A, Ali F (2022) Improving source location privacy in social\ninternet of things using a hybrid phantom routing technique. Comput  Secur 123:102917","DOI":"10.1016\/j.cose.2022.102917"},{"issue":"6","key":"1100_CR4","doi-asserted-by":"publisher","first-page":"734","DOI":"10.3390\/brainsci11060734","volume":"11","author":"T Akter","year":"2021","unstructured":"Akter T, Ali MH, Khan MI, Satu MS, Uddin MJ, Alyami SA, Moni MA (2021) Improved transfer-learning-based facial recognition framework to detect autistic children at an early stage. Brain Sci 11(6):734","journal-title":"Brain Sci"},{"issue":"1","key":"1100_CR5","doi-asserted-by":"publisher","first-page":"18","DOI":"10.37391\/IJEER.100103","volume":"10","author":"M Rahul","year":"2022","unstructured":"Rahul M, Tiwari N, Shukla R, Tyagi D, Yadav V (2022) A new hybrid approach for efficient emotion recognition using deep learning. Int J Electr Electron Res (IJEER) 10(1):18\u201322","journal-title":"Int J Electr Electron Res (IJEER)"},{"issue":"11","key":"1100_CR6","doi-asserted-by":"publisher","first-page":"3993","DOI":"10.3390\/s18113993","volume":"18","author":"M Leo","year":"2018","unstructured":"Leo M, Carcagn\u00ec P, Distante C, Spagnolo P, Mazzeo PL, Rosato AC, Lecciso F (2018) Computational assessment of facial expression production in ASD children. Sensors 18(11):3993","journal-title":"Sensors"},{"key":"1100_CR7","first-page":"1","volume":"2014","author":"S Sun","year":"2014","unstructured":"Sun S, Ge C (2014) A new method of 3D facial expression animation. J Appl Math 2014:1\u20136","journal-title":"J Appl Math"},{"key":"1100_CR8","doi-asserted-by":"crossref","unstructured":"Saste ST, Jagdale SM (2017) Emotion recognition from speech using MFCC and DWT for security system. In: 2017 international conference of electronics, communication and aerospace technology (ICECA)\u00a0(Vol. 1, pp 701\u2013704). IEEE","DOI":"10.1109\/ICECA.2017.8203631"},{"issue":"15","key":"1100_CR9","doi-asserted-by":"publisher","first-page":"12527","DOI":"10.1007\/s00521-022-07292-4","volume":"34","author":"EH Houssein","year":"2022","unstructured":"Houssein EH, Hammad A, Ali AA (2022) Human emotion recognition from EEG-based brain\u2013computer interface using machine learning: a comprehensive review. Neural Comput Appl 34(15):12527\u201312557","journal-title":"Neural Comput Appl"},{"key":"1100_CR10","doi-asserted-by":"publisher","unstructured":"Mollahosseini A, Chan D, Mahoor MH (2016) Going deeper in facial expression recognition using deep neural networks. In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 2016, pp. 1\u201310, https:\/\/doi.org\/10.1109\/WACV.2016.7477450","DOI":"10.1109\/WACV.2016.7477450"},{"key":"1100_CR11","doi-asserted-by":"crossref","unstructured":"Liu P, Han S, Meng Z, Tong Y (2014) Facial expression recognition via a boosted deep belief network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, Columbus, OH, USA,\u00a0pp 1805\u20131812","DOI":"10.1109\/CVPR.2014.233"},{"key":"1100_CR12","doi-asserted-by":"publisher","first-page":"392","DOI":"10.1016\/j.neunet.2022.03.017","volume":"150","author":"P Kumar","year":"2022","unstructured":"Kumar P, Raman B (2022) A BERT based dual-channel explainable text emotion recognition system. Neural Netw 150:392\u2013407","journal-title":"Neural Netw"},{"key":"1100_CR13","doi-asserted-by":"crossref","unstructured":"Han K, Yu D, Tashev I (2014) Speech emotion recognition using deep neural network and extreme learning machine. In: Interspeech 2014","DOI":"10.21437\/Interspeech.2014-57"},{"issue":"2","key":"1100_CR14","first-page":"207","volume":"11","author":"A Hussain","year":"2022","unstructured":"Hussain A, Ahmad M, Hussain T, Ullah I (2022) Efficient content based video retrieval system by applying AlexNet on key frames. ADCAIJ Adv Distrib Comput Artif Intell J 11(2):207\u2013235","journal-title":"ADCAIJ Adv Distrib Comput Artif Intell J"},{"issue":"6","key":"1100_CR15","doi-asserted-by":"publisher","first-page":"6625","DOI":"10.1109\/TPEL.2022.3142113","volume":"37","author":"M Bouzidi","year":"2022","unstructured":"Bouzidi M, Barkat S, Krama A, Abu-Rub H (2022) Generalized predictive direct power control with constant switching frequency for multilevel four-leg grid connected converter. IEEE Trans Power Electron 37(6):6625\u20136636","journal-title":"IEEE Trans Power Electron"},{"key":"1100_CR16","doi-asserted-by":"crossref","unstructured":"Khorrami P, Paine T, Huang T (2015) Do deep neural networks learn facial action units when doing expression recognition. In: Proceedings of the IEEE international conference on computer vision workshops, Santiago, Chile,\u00a0pp 19\u201327","DOI":"10.1109\/ICCVW.2015.12"},{"issue":"8","key":"1100_CR17","doi-asserted-by":"publisher","first-page":"1301","DOI":"10.1109\/JSTSP.2017.2764438","volume":"11","author":"P Tzirakis","year":"2017","unstructured":"Tzirakis P, Trigeorgis G, Nicolaou MA, Schuller BW, Zafeiriou S (2017) End-to-end multimodal emotion recognition using deep neural networks. IEEE J Select Top Signal Process 11(8):1301\u20131309","journal-title":"IEEE J Select Top Signal Process"},{"key":"1100_CR18","doi-asserted-by":"publisher","unstructured":"Zhang W et al. (2022) Transformer-based Multimodal Information Fusion for Facial Expression Analysis. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 2022, pp 2427\u20132436, https:\/\/doi.org\/10.1109\/CVPRW56347.2022.00271","DOI":"10.1109\/CVPRW56347.2022.00271"},{"key":"1100_CR19","doi-asserted-by":"crossref","unstructured":"Zeiler MD, Fergus R (2014) Visualizing and understanding convolutional networks. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part I 13. Springer International Publishing, pp 818\u2013833","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"1100_CR20","doi-asserted-by":"publisher","unstructured":"Bolioli A, Bosca A, Damiano R, Lieto A, Striani M (2022) A complementary account to emotion extraction and classification in cultural heritage based on the Plutchik\u2019s theory. In: Adjunct Proceedings of the 30th ACM Conference on User Modeling, Adaptation and Personalization, New York, NY, United States, pp 374\u2013382. https:\/\/doi.org\/10.1145\/3511047.3537659","DOI":"10.1145\/3511047.3537659"},{"issue":"5","key":"1100_CR21","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0266442","volume":"17","author":"C Correia-Caeiro","year":"2022","unstructured":"Correia-Caeiro C, Burrows A, Wilson DA, Abdelrahman A, Miyabe-Nishiwaki T (2022) CalliFACS: the common marmoset facial action coding system. PLoS ONE 17(5):e0266442","journal-title":"PLoS ONE"},{"key":"1100_CR22","doi-asserted-by":"publisher","unstructured":"Huo Y, Zhang L (2022) OCFER-Net: recognizing facial expression in online learning system. In: Proceedings of the 2022 International Conference on Advanced Visual Interfaces (AVI 2022). Association for Computing Machinery, New York, NY, USA, Article 45, 1\u20133. https:\/\/doi.org\/10.1145\/3531073.3534470","DOI":"10.1145\/3531073.3534470"},{"key":"1100_CR23","doi-asserted-by":"publisher","first-page":"4623188","DOI":"10.1155\/2022\/4623188","volume":"2022","author":"Y Zhang","year":"2022","unstructured":"Zhang Y, Zhang Q, Yang J (2022) Application of an artificial intelligence system recognition based on the deep neural network algorithm. Comput Intell Neurosci 2022:4623188. https:\/\/doi.org\/10.1155\/2022\/4623188","journal-title":"Comput Intell Neurosci"},{"key":"1100_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2022.106621","volume":"215","author":"H Ge","year":"2022","unstructured":"Ge H, Zhu Z, Dai Y, Wang B, Wu X (2022) Facial expression recognition based on deep learning. Comput Methods Programs Biomed 215:106621","journal-title":"Comput Methods Programs Biomed"},{"key":"1100_CR25","doi-asserted-by":"publisher","unstructured":"Karras C, Karras A, Sioutas S (2022) Pattern recognition and event detection on IoT data-streams.\u00a0arXiv preprint. https:\/\/arXiv.org\/2203.01114https:\/\/doi.org\/10.48550\/arXiv.2203.01114","DOI":"10.48550\/arXiv.2203.01114"},{"key":"1100_CR26","unstructured":"Chen J, Chen Z, Chi Z, Fu H (2014) Facial expression recognition based on facial components detection and hog features. In: Scientific Cooperations International Workshops on Electrical and Computer Engineering Subfields 22\u201323 August 2014, Koc University, Istanbul\/Turkey, pp 884\u2013888"},{"key":"1100_CR27","doi-asserted-by":"publisher","unstructured":"Barsoum E, Zhang C, Canton Ferrer C, Zhang Z (2016) Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction (ICMI '16). Association for Computing Machinery, New York, NY, USA, 279\u2013283. https:\/\/doi.org\/10.1145\/2993148.2993165","DOI":"10.1145\/2993148.2993165"},{"key":"1100_CR28","unstructured":"Han S, Meng Z, Khan AS, Tong Y (2016) Incremental boosting convolutional neural network for facial action unit recognition. In: Advances in neural information processing systems,\u00a029"},{"key":"1100_CR29","doi-asserted-by":"publisher","unstructured":"Meng Z, Liu P, Cai J, Han S, Tong Y (2017) Identity-aware convolutional neural network for facial expression recognition. In:\u00a02017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 2017, pp. 558\u2013565, https:\/\/doi.org\/10.1109\/FG.2017.140","DOI":"10.1109\/FG.2017.140"},{"key":"1100_CR30","doi-asserted-by":"publisher","unstructured":"Fernandez P, Pena F, Ren T, Cunha A (2019) FERAtt: facial expression recognition with attention net. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 2019, pp. 837\u2013846. https:\/\/doi.org\/10.1109\/CVPRW.2019.00112","DOI":"10.1109\/CVPRW.2019.00112"},{"key":"1100_CR31","doi-asserted-by":"publisher","unstructured":"Wang K, Peng X, Yang Y, Lu S, Qiao Y (2020) Suppressing uncertainties for large-scale facial expression recognition. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 2020, pp 6896\u20136905, https:\/\/doi.org\/10.1109\/CVPR42600.2020.00693","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"1100_CR32","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang K, Peng X, Yang J, Meng D, Qiao Y (2020) Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans Image Process 29:4057\u20134069","journal-title":"IEEE Trans Image Process"},{"key":"1100_CR33","doi-asserted-by":"publisher","first-page":"1527","DOI":"10.1016\/j.procs.2020.04.163","volume":"171","author":"A Gupta","year":"2020","unstructured":"Gupta A, Arunachalam S, Balakrishnan R (2020) Deep self-attention network for facial emotion recognition. Procedia Comput Sci 171:1527\u20131534","journal-title":"Procedia Comput Sci"},{"key":"1100_CR34","doi-asserted-by":"publisher","first-page":"7383","DOI":"10.1109\/ACCESS.2020.2963913","volume":"8","author":"Y Gan","year":"2020","unstructured":"Gan Y, Chen J, Yang Z, Xu L (2020) Multiple attention network for facial expression recognition. IEEE Access 8:7383\u20137393","journal-title":"IEEE Access"},{"issue":"3","key":"1100_CR35","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2022","unstructured":"Li S, Deng W (2022) Deep facial expression recognition: a survey. IEEE Trans Affect Comput 13(3):1195\u20131215. https:\/\/doi.org\/10.1109\/TAFFC.2020.2981446","journal-title":"IEEE Trans Affect Comput"},{"key":"1100_CR36","unstructured":"Jaderberg M, Simonyan K, Zisserman A (2015) Spatial transformer networks. Adv Neural Inform Process Syst 28"},{"key":"1100_CR37","doi-asserted-by":"publisher","first-page":"7208794","DOI":"10.1155\/2018\/7208794","volume":"2018","author":"X Wang","year":"2018","unstructured":"Wang X, Wang X, Ni Y (2018) Unsupervised domain adaptation for facial expression recognition using generative adversarial networks. Comput Intell Neurosci 2018:7208794. https:\/\/doi.org\/10.1155\/2018\/7208794","journal-title":"Comput Intell Neurosci"},{"issue":"4","key":"1100_CR38","first-page":"123","volume":"31","author":"M Khalid","year":"2021","unstructured":"Khalid M, Keming M, Hussain T (2021) Design and implementation of clothing fashion style recommendation system using deep learning. Rom J Inform Technol Autom Control 31(4):123\u2013136","journal-title":"Rom J Inform Technol Autom Control"},{"key":"1100_CR39","doi-asserted-by":"publisher","first-page":"64827","DOI":"10.1109\/ACCESS.2019.2917266","volume":"7","author":"MI Georgescu","year":"2019","unstructured":"Georgescu MI, Ionescu RT, Popescu M (2019) Local learning with deep and handcrafted features for facial expression recognition. IEEE Access 7:64827\u201364836","journal-title":"IEEE Access"},{"key":"1100_CR40","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-66790-4_1","volume-title":"Advances in hybridization of intelligent methods smart innovation, systems and technologies","author":"P Giannopoulos","year":"2018","unstructured":"Giannopoulos P, Perikos I, Hatzilygeroudis I (2018) Deep learning approaches for facial emotion recognition: a case study on FER-2013. In: Hatzilygeroudis I, Palade V (eds) Advances in hybridization of intelligent methods smart innovation, systems and technologies, vol 85. Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-319-66790-4_1"},{"issue":"9","key":"1100_CR41","doi-asserted-by":"publisher","first-page":"3046","DOI":"10.3390\/s21093046","volume":"21","author":"S Minaee","year":"2021","unstructured":"Minaee S, Minaei M, Abdolrashidi A (2021) Deep-emotion: facial expression recognition using attentional convolutional network. Sensors 21(9):3046","journal-title":"Sensors"},{"key":"1100_CR42","doi-asserted-by":"publisher","unstructured":"Kollias D, Zafeiriou S (2019) Expression, affect, action unit recognition: Aff-wild2, multi-task learning and arcface.\u00a0arXiv preprint. https:\/\/arXiv.org\/1910.04855https:\/\/doi.org\/10.48550\/arXiv.1910.04855","DOI":"10.48550\/arXiv.1910.04855"},{"key":"1100_CR43","doi-asserted-by":"publisher","first-page":"8828245","DOI":"10.1155\/2021\/8828245","volume":"2021","author":"B Niu","year":"2021","unstructured":"Niu B, Gao Z, Guo B (2021) Facial expression recognition with LBP and ORB features. Comput Intell Neurosci 2021:8828245. https:\/\/doi.org\/10.1155\/2021\/8828245","journal-title":"Comput Intell Neurosci"},{"key":"1100_CR44","doi-asserted-by":"publisher","first-page":"38151","DOI":"10.1007\/s11042-022-13169-4","volume":"81","author":"Irfanullah","year":"2022","unstructured":"Irfanullah, Hussain T, Iqbal A et al (2022) Real time violence detection in surveillance videos using convolutional neural networks. Multimed Tools Appl 81:38151\u201338173. https:\/\/doi.org\/10.1007\/s11042-022-13169-4","journal-title":"Multimed Tools Appl"},{"issue":"5","key":"1100_CR45","doi-asserted-by":"publisher","first-page":"3211","DOI":"10.1007\/s11227-018-2554-8","volume":"76","author":"H Wang","year":"2020","unstructured":"Wang H, Wei S, Fang B (2020) Facial expression recognition using iterative fusion of MO-HOG and deep features. J Supercomput 76(5):3211\u20133221","journal-title":"J Supercomput"},{"key":"1100_CR46","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1007\/978-981-16-4369-9_59","volume-title":"Advanced computational paradigms and hybrid intelligent computing","author":"S Bhattacharya","year":"2022","unstructured":"Bhattacharya S (2022) A survey on: facial expression recognition using various deep learning techniques. Advanced computational paradigms and hybrid intelligent computing. Springer, Singapore, pp 619\u2013631"},{"key":"1100_CR47","doi-asserted-by":"publisher","unstructured":"Shima Y, Omori Y (2018) Image augmentation for classifying facial expression images by using deep neural network pre-trained with object image database. In: Proceedings of the 3rd International Conference on Robotics, Control and Automation (ICRCA\u201918). Association for Computing Machinery, New York, NY, USA, 140\u2013146. https:\/\/doi.org\/10.1145\/3265639.3265664","DOI":"10.1145\/3265639.3265664"},{"issue":"2s","key":"1100_CR48","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3341095","volume":"16","author":"SH Wang","year":"2020","unstructured":"Wang SH, Zhang YD (2020) DenseNet-201-based deep neural network with composite learning factor and precomputation for multiple sclerosis classification. ACM Trans Multimed Comput Commun Appl (TOMM) 16(2s):1\u201319","journal-title":"ACM Trans Multimed Comput Commun Appl (TOMM)"},{"issue":"4","key":"1100_CR49","doi-asserted-by":"publisher","first-page":"687","DOI":"10.3390\/sym14040687","volume":"14","author":"K Zaman","year":"2022","unstructured":"Zaman K, Sun Z, Shah SM, Shoaib M, Pei L, Hussain A (2022) Driver Emotions recognition based on improved faster R-CNN and neural architectural search network. Symmetry 14(4):687","journal-title":"Symmetry"},{"key":"1100_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102121","volume":"72","author":"A Sohail","year":"2021","unstructured":"Sohail A, Khan A, Nisar H, Tabassum S, Zameer A (2021) Mitotic nuclei analysis in breast cancer histopathology images using deep ensemble classifier. Med Image Anal 72:102121","journal-title":"Med Image Anal"},{"key":"1100_CR51","doi-asserted-by":"crossref","unstructured":"Hong S, Wu M, Zhou Y, Wang Q, Shang J, Li H, Xie J (2017) ENCASE: an ENsemble ClASsifiEr for ECG classification using expert features and deep neural networks. In: 2017 Computing in cardiology (cinc). IEEE, pp. 1\u20134","DOI":"10.22489\/CinC.2017.178-245"},{"issue":"3","key":"1100_CR52","doi-asserted-by":"publisher","first-page":"345","DOI":"10.1109\/LGRS.2017.2787421","volume":"15","author":"X Gu","year":"2018","unstructured":"Gu X, Angelov PP, Zhang C, Atkinson PM (2018) A massively parallel deep rule-based ensemble classifier for remote sensing scenes. IEEE Geosci Remote Sens Lett 15(3):345\u2013349","journal-title":"IEEE Geosci Remote Sens Lett"},{"key":"1100_CR53","doi-asserted-by":"publisher","unstructured":"Potamias RA, Siolas G, Stafylopatis A (2019) A robust deep ensemble classifier for figurative language detection. In: Macintyre J, Iliadis L, Maglogiannis I, Jayne C (eds) Engineering applications of neural networks. EANN 2019. Communications in Computer and Information Science, vol 1000. Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-030-20257-6_14","DOI":"10.1007\/978-3-030-20257-6_14"},{"key":"1100_CR54","doi-asserted-by":"publisher","first-page":"9579","DOI":"10.1007\/s00521-020-04842-6","volume":"34","author":"PM Shakeel","year":"2022","unstructured":"Shakeel PM, Burhanuddin MA, Desa MI (2022) Automatic lung cancer detection from CT image using improved deep neural network and ensemble classifier. Neural Comput Appl 34:9579\u20139592. https:\/\/doi.org\/10.1007\/s00521-020-04842-6","journal-title":"Neural Comput Appl"},{"issue":"10","key":"1100_CR55","doi-asserted-by":"publisher","first-page":"3959","DOI":"10.3390\/s22103959","volume":"22","author":"SM Shah","year":"2022","unstructured":"Shah SM, Sun Z, Zaman K, Hussain A, Shoaib M, Pei L (2022) A driver gaze estimation method based on deep learning. Sensors 22(10):3959","journal-title":"Sensors"},{"issue":"15","key":"1100_CR56","doi-asserted-by":"publisher","first-page":"2281","DOI":"10.3390\/electronics11152281","volume":"11","author":"R Ullah","year":"2022","unstructured":"Ullah R, Gani A, Shiraz M, Yousufzai IK, Zaman K (2022) Auction mechanism-based sectored fractional frequency reuse for irregular geometry multicellular networks. Electronics 11(15):2281","journal-title":"Electronics"},{"issue":"7","key":"1100_CR57","doi-asserted-by":"publisher","first-page":"2533","DOI":"10.1007\/s00521-018-3937-8","volume":"32","author":"S Malakar","year":"2020","unstructured":"Malakar S, Ghosh M, Bhowmik S, Sarkar R, Nasipuri M (2020) A GA based hierarchical feature selection approach for handwritten word recognition. Neural Comput Appl 32(7):2533\u20132552","journal-title":"Neural Comput Appl"},{"issue":"21","key":"1100_CR58","doi-asserted-by":"publisher","first-page":"2705","DOI":"10.3390\/math9212705","volume":"9","author":"N Bacanin","year":"2021","unstructured":"Bacanin N, Stoean R, Zivkovic M, Petrovic A, Rashid TA, Bezdan T (2021) Performance of a novel chaotic firefly algorithm with enhanced exploration for tackling global optimization problems: application for dropout regularization. Mathematics 9(21):2705","journal-title":"Mathematics"},{"key":"1100_CR59","doi-asserted-by":"publisher","unstructured":"Lyons MJ, Akamatsu S, Kamachi M, Gyoba J (1998) Coding facial expressions with Gabor wavelets. In: 3rd IEEE International Conference on Automatic Face and Gesture Recognition, pp 200\u2013205. https:\/\/doi.org\/10.1109\/AFGR.1998.670949. Open access content available at: https:\/\/zenodo.org\/record\/3430156","DOI":"10.1109\/AFGR.1998.670949"}],"updated-by":[{"DOI":"10.1007\/s40747-023-01338-3","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2024,1,19]],"date-time":"2024-01-19T00:00:00Z","timestamp":1705622400000}},{"DOI":"10.1007\/s40747-023-01195-0","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2023,8,14]],"date-time":"2023-08-14T00:00:00Z","timestamp":1691971200000}}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01100-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-023-01100-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01100-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,20]],"date-time":"2024-01-20T07:34:19Z","timestamp":1705736059000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-023-01100-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,7]]},"references-count":59,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["1100"],"URL":"https:\/\/doi.org\/10.1007\/s40747-023-01100-9","relation":{"correction":[{"id-type":"doi","id":"10.1007\/s40747-023-01338-3","asserted-by":"object"},{"id-type":"doi","id":"10.1007\/s40747-023-01195-0","asserted-by":"object"}]},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,7]]},"assertion":[{"value":"5 November 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 May 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 June 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 August 2023","order":4,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":5,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":6,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s40747-023-01195-0","URL":"https:\/\/doi.org\/10.1007\/s40747-023-01195-0","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 January 2024","order":8,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":9,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":10,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s40747-023-01338-3","URL":"https:\/\/doi.org\/10.1007\/s40747-023-01338-3","order":11,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All the authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}