{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,27]],"date-time":"2026-04-27T10:08:41Z","timestamp":1777284521888,"version":"3.51.4"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T00:00:00Z","timestamp":1667433600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T00:00:00Z","timestamp":1667433600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Xiamen University Malaysia Research Fund","award":["XMUMRF\/2022- C9\/IECE\/0035"],"award-info":[{"award-number":["XMUMRF\/2022- C9\/IECE\/0035"]}]},{"name":"Taif University Researchers Supporting Project Number","award":["TURSP-2020\/79"],"award-info":[{"award-number":["TURSP-2020\/79"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cluster Comput"],"published-print":{"date-parts":[[2023,4]]},"DOI":"10.1007\/s10586-022-03705-0","type":"journal-article","created":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T14:02:42Z","timestamp":1667484162000},"page":"1253-1266","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":27,"title":["An AI-empowered affect recognition model for healthcare and emotional well-being using physiological signals"],"prefix":"10.1007","volume":"26","author":[{"given":"Zijian","family":"Zhou","sequence":"first","affiliation":[]},{"given":"Muhammad Adeel","family":"Asghar","sequence":"additional","affiliation":[]},{"given":"Daniyal","family":"Nazir","sequence":"additional","affiliation":[]},{"given":"Kamran","family":"Siddique","sequence":"additional","affiliation":[]},{"given":"Mohammad","family":"Shorfuzzaman","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2284-0479","authenticated-orcid":false,"given":"Raja Majid","family":"Mehmood","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,3]]},"reference":[{"issue":"19","key":"3705_CR1","doi-asserted-by":"publisher","first-page":"8402","DOI":"10.1109\/JSEN.2018.2867221","volume":"19","author":"A Albraikan","year":"2019","unstructured":"Albraikan, A., Tobin, D.P., El Saddik, A.: Toward user-independent emotion recognition using physiological signals. IEEE Sens. J. 19(19), 8402\u20138412 (2019)","journal-title":"IEEE Sens. J."},{"issue":"5","key":"3705_CR2","doi-asserted-by":"publisher","first-page":"565","DOI":"10.1007\/s00530-017-0561-x","volume":"25","author":"MS Hossain","year":"2019","unstructured":"Hossain, M.S., Muhammad, G., Alamri, A.: Smart healthcare monitoring: a voice pathology detection paradigm for smart cities. Multimed. Syst. 25(5), 565\u2013575 (2019)","journal-title":"Multimed. Syst."},{"issue":"6","key":"3705_CR3","doi-asserted-by":"publisher","first-page":"120","DOI":"10.1109\/MNET.011.2000064","volume":"34","author":"MS Hossain","year":"2020","unstructured":"Hossain, M.S., Muhammad, G.: Deep learning based pathology detection for smart connected healthcare. IEEE Netw. 34(6), 120\u2013125 (2020)","journal-title":"IEEE Netw."},{"issue":"4","key":"3705_CR4","doi-asserted-by":"publisher","first-page":"1751","DOI":"10.1109\/TCSVT.2021.3080928","volume":"32","author":"Fangzheng Tian","year":"2022","unstructured":"Tian, Fangzheng, Gao, Yongbin, Fang, Zhijun, Fang, Yuming, Jia, Gu., Fugita, Hamido, Hwang, Jenq-Neng.: Depth estimation using a self-supervised network based on cross-layer feature fusion and the quadtree constraint. IEEE Trans. Circuits Syst. Video Technol. (TCSVT) 32(4), 1751\u20131766 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol. (TCSVT)"},{"issue":"2","key":"3705_CR5","doi-asserted-by":"publisher","first-page":"164","DOI":"10.3390\/electronics8020164","volume":"8","author":"Yeong-Seok Seo","year":"2019","unstructured":"Seo, Yeong-Seok., Huh, Jun-Ho.: Automatic emotion-based music classification for supporting intelligent IoT applications. Electronics 8(2), 164 (2019)","journal-title":"Electronics"},{"key":"3705_CR6","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.entcs.2019.04.009","volume":"343","author":"E Maria","year":"2019","unstructured":"Maria, E., Matthias, L., Sten, H.: Emotion recognition from physiological signal analysis: a review. Electron. Notes Theor. Comput. Scie. 343, 35\u201355 (2019)","journal-title":"Electron. Notes Theor. Comput. Scie."},{"issue":"592","key":"3705_CR7","first-page":"162","volume":"20","author":"A Dzedzickis","year":"2020","unstructured":"Dzedzickis, A., Kaklauska, C., Bucinskas, C.: Human emotion recognition: review of sensors and method. Sensors 20(592), 162\u2013186 (2020)","journal-title":"Sensors"},{"key":"3705_CR8","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1016\/j.eswa.2017.09.062","volume":"93","author":"B Nakisa","year":"2018","unstructured":"Nakisa, B., Rastgoo, M.N., Tjondronegoro, D.: Evolutionary computation algorithms for feature selection of EEG-based emotion recognition using mobile sensors. Expert Syst. Appl. 93, 143\u2013155 (2018)","journal-title":"Expert Syst. Appl."},{"issue":"6","key":"3705_CR9","doi-asserted-by":"publisher","first-page":"2266","DOI":"10.1109\/JSEN.2018.2883497","volume":"19","author":"V Gupta","year":"2019","unstructured":"Gupta, V., Chopda, M.D., Pachori, R.B.: Cross-subject emotion recognition using flexible analytic wavelet transform from EEG signals. IEEE Sens. J. 19(6), 2266\u20132274 (2019)","journal-title":"IEEE Sens. J."},{"key":"3705_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107700","volume":"113","author":"M Shorfuzzaman","year":"2021","unstructured":"Shorfuzzaman, M., Hossain, M.S.: MetaCOVID: a siamese neural network framework with contrastive loss for n-shot diagnosis of COVID-19 patients. Pattern Recognit. 113, 107700 (2021). https:\/\/doi.org\/10.1016\/j.patcog.2020.107700","journal-title":"Pattern Recognit."},{"key":"3705_CR11","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.inffus.2018.09.008","volume":"49","author":"MS Hossain","year":"2019","unstructured":"Hossain, M.S., Muhammad, G.: Emotion recognition using deep learning approach from audio-visual emotional big data. Inf. Fusion 49, 69\u201378 (2019)","journal-title":"Inf. Fusion"},{"issue":"2","key":"3705_CR12","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1109\/JSAC.2020.3020654","volume":"39","author":"G Muhammad","year":"2020","unstructured":"Muhammad, G., Hossain, M.S., Kumar, N.: EEG-based pathology detection for home health monitoring. IEEE J. Sel. Areas Commun. 39(2), 603\u2013610 (2020)","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"3705_CR13","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Ghosh, S.: From individual to group-level emotion recognition: Emotiw 5.0. In: Proceedings of the 19th ACM international conference on multimodal interaction, pp. 524\u2013528. (2017)","DOI":"10.1145\/3136755.3143004"},{"key":"3705_CR14","doi-asserted-by":"crossref","unstructured":"Fan, Y., Lu, X., Li, D., Liu, Y.: Video-based emotion recognition using CNN-RNN and C3D hybrid networks. In: Proceedings of the 18th ACM international conference on multimodal interaction, pp. 445\u2013450. (2016)","DOI":"10.1145\/2993148.2997632"},{"key":"3705_CR15","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.neunet.2017.02.013","volume":"92","author":"H Fayek","year":"2017","unstructured":"Fayek, H., Lech, M., Cavedon, L.: Evaluating deep learning architectures for speech emotion recognition. Neural Netw. 92, 60\u201368 (2017)","journal-title":"Neural Netw."},{"issue":"2","key":"3705_CR16","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1007\/s12193-015-0195-2","volume":"10","author":"S Kahou","year":"2016","unstructured":"Kahou, S., Bouthillier, X., Lamblin, P.: Emonets: multimodal deep learning approaches for emotion recognition in video. J. Multimodal User Interfaces 10(2), 99\u2013111 (2016)","journal-title":"J. Multimodal User Interfaces"},{"key":"3705_CR17","doi-asserted-by":"crossref","unstructured":"Mirsamadi, S., Barsoum, E., Zhang, C.: Automatic speech emotion recognition using recurrent neural networks with local attention. In: 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP). pp. 2227\u20132231. IEEE (2017)","DOI":"10.1109\/ICASSP.2017.7952552"},{"key":"3705_CR18","doi-asserted-by":"crossref","unstructured":"Poria, S., Chaturvedi, I., Cambria, E., Hussain, H.: A convolutional MKL based multimodal emotion recognition and sentiment analysis. In: 2016 IEEE 16th international conference on data mining (ICDM). pp 439\u2013448. IEEE (2016)","DOI":"10.1109\/ICDM.2016.0055"},{"key":"3705_CR19","doi-asserted-by":"crossref","unstructured":"Trigeorgis, G., Ringeval, F., Brueckner, B.: End-to-end speech emotion recognition using a deep convolutional recurrent network. In: 2016 IEEE international conference on acoustics, speech and signal processing, pp. 5200\u20135204. (2016)","DOI":"10.1109\/ICASSP.2016.7472669"},{"issue":"4","key":"3705_CR20","doi-asserted-by":"publisher","first-page":"2399","DOI":"10.1109\/JIOT.2017.2772959","volume":"5","author":"MS Hossain","year":"2018","unstructured":"Hossain, M.S., Muhammad, G.: Emotion-aware connected healthcare big data towards 5G. IEEE Internet Things J. 5(4), 2399\u20132406 (2018)","journal-title":"IEEE Internet Things J."},{"issue":"4","key":"3705_CR21","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/JSEN.2018.2873402","volume":"19","author":"B Geethanjali","year":"2018","unstructured":"Geethanjali, B., Adalarasu, K., Jagannath, M., Guhan Seshadri, N.P.: Music-induced brain functional connectivity using EEG sensors: a study on Indian music. IEEE Sens. J. 19(4), 1499\u20131507 (2018)","journal-title":"IEEE Sens. J."},{"issue":"3","key":"3705_CR22","doi-asserted-by":"publisher","first-page":"716","DOI":"10.1109\/JSEN.2016.2623677","volume":"17","author":"A Greco","year":"2016","unstructured":"Greco, A., Valenza, G., Citi, L., Scilingo, E.P.: Arousal and valence recognition of affective sounds based on electrodermal activity. IEEE Sens. J. 17(3), 716\u2013725 (2016)","journal-title":"IEEE Sens. J."},{"issue":"8","key":"3705_CR23","doi-asserted-by":"publisher","first-page":"1301","DOI":"10.1109\/JSTSP.2017.2764438","volume":"11","author":"P Tzirakis","year":"2017","unstructured":"Tzirakis, P., Trigeorgis, G., Nicolaou, M.A., Schuller, B.W., Zafeiriou, S.: End-to-end multimodal emotion recognition using deep neural networks. IEEE J. Sel. Top. Signal Process. 11(8), 1301\u20131309 (2017)","journal-title":"IEEE J. Sel. Top. Signal Process."},{"issue":"1","key":"3705_CR24","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","volume":"3","author":"S Koelstra","year":"2012","unstructured":"Koelstra, S., et al.: DEAP: a database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 3(1), 18\u201331 (2012)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"1","key":"3705_CR25","doi-asserted-by":"publisher","first-page":"149","DOI":"10.1007\/s11031-017-9648-0","volume":"42","author":"R Wrigh","year":"2016","unstructured":"Wrigh, R., Riedel, R., Sechres, L.: Sex differences in emotion recognition ability: the mediating role of trait emotional awareness. Motiv. Emot. 42(1), 149\u2013160 (2016)","journal-title":"Motiv. Emot."},{"key":"3705_CR26","first-page":"1","volume":"3","author":"Z Liu","year":"2016","unstructured":"Liu, Z., Xie, Q., Li, S., et al.: Electroencephalogram emotion recognition based on empirical mode decomposition and optimal feature selection. IEEE Trans. Cognit. Dev. Syst. 3, 1\u20134 (2016)","journal-title":"IEEE Trans. Cognit. Dev. Syst."},{"key":"3705_CR27","doi-asserted-by":"crossref","unstructured":"Cao, G., Ma, Y., Meng, X. et al.: Emotion recognition based on CNN. In: 2019 Chinese Control Conference (CCC), vol. 2, pp. 8627\u20138630. (2019)","DOI":"10.23919\/ChiCC.2019.8866540"},{"key":"3705_CR28","doi-asserted-by":"crossref","unstructured":"Han, B., Lee, S.: Feature selection and comparison for the emotion recognition according to music listening. In: 2017 international conference on robotics and automation sciences, vol. 7, pp. 172\u2013176. (2017)","DOI":"10.1109\/ICRAS.2017.8071939"},{"key":"3705_CR29","doi-asserted-by":"crossref","unstructured":"Shahnaz, C., Shoaib, M.: Emotion recognition based on wavelet analysis of empirical mode decomposed EEG signals responsive to music videos. In: 2016 IEEE region 10 conference, vol. 11, pp. 424\u2013427. (2016)","DOI":"10.1109\/TENCON.2016.7848034"},{"key":"3705_CR30","doi-asserted-by":"crossref","unstructured":"Li, X., Song, D., Zhang, P. et al.: Emotion recognition from multi-channel EEG data through convolutional recurrent neural network. In: 2016 IEEE international conference on bioinformatics and biomedicine, vol. 7, pp. 352\u2013359. (2016)","DOI":"10.1109\/BIBM.2016.7822545"},{"key":"3705_CR31","doi-asserted-by":"crossref","unstructured":"Islam, R., Ahmad, M.: Wavelet analysis based classification of emotion from EEG signal. In: 2019 international conference on electrical, computer and communication engineering (ECCE), pp. 7\u20139. (2019)","DOI":"10.1109\/ECACE.2019.8679156"},{"key":"3705_CR32","doi-asserted-by":"crossref","unstructured":"Shao, J., Zhu, J., Wei, Y. et al.: Emotion recognition by edge-weighted hypergraph neural network. In: 2019 IEEE international conference on image processing, vol. 12, pp. 425\u2013431. (2019)","DOI":"10.1109\/ICIP.2019.8803207"}],"container-title":["Cluster Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-022-03705-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10586-022-03705-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-022-03705-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,9]],"date-time":"2025-04-09T14:18:00Z","timestamp":1744208280000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10586-022-03705-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,11,3]]},"references-count":32,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,4]]}},"alternative-id":["3705"],"URL":"https:\/\/doi.org\/10.1007\/s10586-022-03705-0","relation":{},"ISSN":["1386-7857","1573-7543"],"issn-type":[{"value":"1386-7857","type":"print"},{"value":"1573-7543","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,11,3]]},"assertion":[{"value":"23 February 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 June 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 August 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 November 2022","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or non-financial interests to disclose","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Ethics approval was not required for this research.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"All authors signed the consent form.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}},{"value":"This work did not involve humans and animals.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Research involving human and animal participants"}}]}}