{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T14:24:05Z","timestamp":1761402245837,"version":"build-2065373602"},"reference-count":42,"publisher":"MDPI AG","issue":"14","license":[{"start":{"date-parts":[[2022,7,9]],"date-time":"2022-07-09T00:00:00Z","timestamp":1657324800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Ministry of Science and Technology, Republic of China","award":["110-2221-E-027-054-MY3"],"award-info":[{"award-number":["110-2221-E-027-054-MY3"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Human emotions are variant with time, non-stationary, complex in nature, and are invoked as a result of human reactions during our daily lives. Continuously detecting human emotions from one-dimensional EEG signals is an arduous task. This paper proposes an advanced signal processing mechanism for emotion detection from EEG signals using continuous wavelet transform. The space and time components of the raw EEG signals are converted into 2D spectrograms followed by feature extraction. A hybrid spatio-temporal deep neural network is implemented to extract rich features. A differential-based entropy feature selection technique adaptively differentiates features based on entropy, based on low and high information regions. Bag of Deep Features (BoDF) is applied to create clusters of similar features and computes the features vocabularies for reduction of feature dimensionality. Extensive experiments are performed on the SEED dataset, which shows the significance of the proposed method compared to state-of-the-art methods. Specifically, the proposed model achieved 96.7%, 96.2%, 95.8%, and 95.3% accuracy with the SJTU SEED dataset, for SVM, ensemble, tree, and KNN classifiers, respectively.<\/jats:p>","DOI":"10.3390\/s22145158","type":"journal-article","created":{"date-parts":[[2022,7,11]],"date-time":"2022-07-11T00:06:21Z","timestamp":1657497981000},"page":"5158","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["A Hybrid Hand-Crafted and Deep Neural Spatio-Temporal EEG Features Clustering Framework for Precise Emotional Status Recognition"],"prefix":"10.3390","volume":"22","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1302-7689","authenticated-orcid":false,"given":"Qazi Mazhar ul","family":"Haq","sequence":"first","affiliation":[{"name":"Department of Electrical Engineering, National Taipei University of Technology, Taipei 106, Taiwan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0128-575X","authenticated-orcid":false,"given":"Leehter","family":"Yao","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering, National Taipei University of Technology, Taipei 106, Taiwan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6902-5455","authenticated-orcid":false,"given":"Wahyu","family":"Rahmaniar","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering, National Taipei University of Technology, Taipei 106, Taiwan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3860-2635","authenticated-orcid":false,"family":"Fawad","sequence":"additional","affiliation":[{"name":"Department of Telecommunication Engineering, University of Engineering and Technology Taxila, Rawalpindi 47050, Punjab, Pakistan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4754-3320","authenticated-orcid":false,"given":"Faizul","family":"Islam","sequence":"additional","affiliation":[{"name":"School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China"}]}],"member":"1968","published-online":{"date-parts":[[2022,7,9]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"1068","DOI":"10.1037\/a0023749","article-title":"Misery loves company: Mood-congruent emotional responding to music","volume":"11","author":"Hunter","year":"2011","journal-title":"Emotion"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"1075","DOI":"10.1109\/TMM.2014.2305573","article-title":"Corpus development for affective video indexing","volume":"16","author":"Soleymani","year":"2014","journal-title":"IEEE Trans. Multimed."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"505","DOI":"10.1007\/s11042-010-0632-x","article-title":"Looking at the viewer: Analysing facial activity to detect personal highlights of multimedia contents","volume":"51","author":"Joho","year":"2010","journal-title":"Multimed. Tools Appl."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"669","DOI":"10.1038\/372669a0","article-title":"Impaired recognition of emotion in facial expressions following bilateral damage to the human amygdala","volume":"372","author":"Adolphs","year":"1994","journal-title":"Nature"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"151","DOI":"10.1016\/j.heares.2014.10.003","article-title":"Voice emotion recognition by cochlear-implanted children and their normally-hearing peers","volume":"322","author":"Chatterjee","year":"2015","journal-title":"Hear. Res."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Ross, P.D., Polson, L., and Grosbras, M.H. (2012). Developmental Changes in Emotion Recognition from Full-Light and Point-Light Displays of Body Movement. PLoS ONE, 7.","DOI":"10.1371\/journal.pone.0044815"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1016\/j.cmpb.2016.12.005","article-title":"Recognition of emotions using multimodal physiological signals and an ensemble deep learning model","volume":"140","author":"Zhong","year":"2017","journal-title":"Comput. Meth. Prog. Bio."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Abadi, M.K., Kia, M., Subramanian, R., Avesani, P., and Sebe, N. (2013, January 22\u201326). Decoding affect in videos employing the MEG brain signal. Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China.","DOI":"10.1109\/FG.2013.6553809"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"197","DOI":"10.1177\/1557234X11410385","article-title":"Affective interaction: Understanding, evaluating, and designing for human emotion","volume":"7","author":"Lottridge","year":"2011","journal-title":"Rev. Hum. Factors Ergon."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"847","DOI":"10.1037\/0022-3514.64.5.847","article-title":"Stability of emotion experiences and their relations to traits of personality","volume":"64","author":"Izard","year":"1993","journal-title":"J. Personal. Soc. Psychol."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"372","DOI":"10.1037\/0003-066X.50.5.372","article-title":"The emotion probe: Studies of motivation and attention","volume":"50","author":"Lang","year":"1995","journal-title":"Am. Psychol."},{"key":"ref_12","unstructured":"Ekman, P.E., and Davidson, R.J. (1994). The Nature of Emotion: Fundamental Questions, Oxford University Press."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"38","DOI":"10.1147\/sj.421.0038","article-title":"Affect and machine design: Lessons for the development of autonomous machines","volume":"42","author":"Norman","year":"2003","journal-title":"IBM Syst. J."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"839","DOI":"10.1109\/TCYB.2017.2788081","article-title":"Spatial-Temporal Recurrent Neural Network for Emotion Recognition","volume":"49","author":"Zhang","year":"2019","journal-title":"IEEE Trans. Cybern."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"162","DOI":"10.3389\/fnins.2018.00162","article-title":"Exploring EEG features in cross-subject emotion recognition","volume":"12","author":"Li","year":"2018","journal-title":"Front. Neurosci."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"08TR02","DOI":"10.1088\/1361-6579\/aad57e","article-title":"Passive BCI beyond the lab: Current trends and future directions","volume":"39","author":"Borghini","year":"2018","journal-title":"Physiol. Meas."},{"key":"ref_17","unstructured":"Keltiner, D., Lewis, P.E.M., and Jones, J.M.H. (2000). Facial Expression of Emotion, Hand Book of Emotions, Gilford Press."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1016\/j.artmed.2013.07.006","article-title":"Asynchronous gaze-independent event-related potential-based brain\u2013Computer interface","volume":"59","author":"Aloise","year":"2013","journal-title":"Artif. Intell. Med."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1007\/s10462-012-9368-5","article-title":"Features and classifiers for emotion recognition from speech: A survey from 2000 to 2011","volume":"43","author":"Anagnostopoulos","year":"2012","journal-title":"Artif. Intell. Rev."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"355","DOI":"10.1109\/TBME.2010.2082539","article-title":"Regularizing Common Spatial Patterns to Improve BCI Designs: Unified Theory and New Algorithms","volume":"58","author":"Lotte","year":"2011","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"10840","DOI":"10.1109\/ACCESS.2018.2809453","article-title":"Brain-Computer Interface Control in a Virtual Reality Environment and Applications for the Internet of Things","volume":"6","author":"Coogan","year":"2018","journal-title":"IEEE Access"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"057180","DOI":"10.1155\/2007\/57180","article-title":"Classifying EEG for Brain-Computer Interface: Learning Optimal Filters for Dynamical System Features","volume":"2007","author":"Song","year":"2007","journal-title":"Comput. Intell. Neurosci."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1007\/s10044-016-0567-6","article-title":"Emotion recognition from EEG signals by using multivariate empirical mode decomposition","volume":"21","author":"Mert","year":"2016","journal-title":"Pattern Anal. Appl."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Chondro, P., Haq, Q.M.U., Ruan, S.J., and Li, L.P.H. (2020). Transferable architecture for segmenting maxillary sinuses on texture-enhanced occipitomental view radiographs. Mathematics, 8.","DOI":"10.3390\/math8050768"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"127678","DOI":"10.1109\/ACCESS.2019.2939623","article-title":"Motor Imagery EEG Signals Classification Based on Mode Amplitude and Frequency Components Using Empirical Wavelet Transform","volume":"7","author":"Sadiq","year":"2019","journal-title":"IEEE Access"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"398","DOI":"10.1016\/j.bspc.2016.09.007","article-title":"Comparison of signal decomposition methods in classification of EEG signals for motor-imagery BCI system","volume":"31","author":"Kevric","year":"2017","journal-title":"Biomed. Signal Process. Control"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Raheel, A., Majid, M., and Anwar, S.M. (2019, January 30\u201331). Facial expression recognition based on electroencephalography. Proceedings of the 2019 2nd International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan.","DOI":"10.1109\/ICOMET.2019.8673408"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Nguyen, B.T., Trinh, M.H., Phan, T.V., and Nguyen, H.D. (2017, January 16\u201319). An efficient real-time emotion detection using camera and facial landmarks. Proceedings of the 2017 Seventh International Conference on Information Science and Technology (ICIST), Da Nang, Vietnam.","DOI":"10.1109\/ICIST.2017.7926765"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Chao, H., Dong, L., Liu, Y., and Lu, B. (2019). Emotion Recognition from Multiband EEG Signals Using CapsNet. Sensors, 19.","DOI":"10.3390\/s19092212"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.cmpb.2018.04.005","article-title":"Deep learning for healthcare applications based on physiological signals: A review","volume":"161","author":"Faust","year":"2018","journal-title":"Comput. Methods Programs Biomed."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"2266","DOI":"10.1109\/JSEN.2018.2883497","article-title":"Cross-Subject Emotion Recognition Using Flexible Analytic Wavelet Transform From EEG Signals","volume":"19","author":"Gupta","year":"2019","journal-title":"IEEE Sens. J."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"1093","DOI":"10.1016\/j.asej.2014.04.011","article-title":"Sentiment analysis algorithms and applications: A survey","volume":"5","author":"Medhat","year":"2014","journal-title":"Ain Shams Eng. J."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"617","DOI":"10.1007\/s10115-018-1236-4","article-title":"A survey of sentiment analysis in social media","volume":"60","author":"Yue","year":"2019","journal-title":"Knowl. Inf. Syst."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Gore, R.J., Diallo, S., and Padilla, J. (2015). You are what you tweet: Connecting the geographic variation in America\u2019s obesity rate to twitter content. PLoS ONE, 10.","DOI":"10.1371\/journal.pone.0133505"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Padilla, J.J., Kavak, H., Lynch, C.J., Gore, R.J., and Diallo, S.Y. (2018). Temporal and spatiotemporal investigation of tourist attraction visit sentiment on Twitter. PLoS ONE, 13.","DOI":"10.1371\/journal.pone.0198857"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Mitchell, L., Frank, M.R., Harris, K.D., Dodds, P.S., and Danforth, C.M. (2013). The geography of happiness: Connecting twitter sentiment and expression, demographics, and objective characteristics of place. PLoS ONE, 8.","DOI":"10.1371\/journal.pone.0064417"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"2625","DOI":"10.1038\/srep02625","article-title":"Happiness and the patterns of life: A study of geolocated tweets","volume":"3","author":"Frank","year":"2013","journal-title":"Sci. Rep."},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Asghar, M.A., Khan, M.J., Amin, Y., Rizwan, M., Rahman, M., Badnava, S., and Mirjavadi, S.S. (2019). EEG-Based Multi-Modal Emotion Recognition using Bag of Deep Features: An Optimal Feature Selection Approach. Sensors, 19.","DOI":"10.3390\/s19235218"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"143","DOI":"10.1016\/j.eswa.2017.09.062","article-title":"Evolutionary computation algorithms for feature selection of EEG-based emotion recognition using mobile sensors","volume":"93","author":"Nakisa","year":"2018","journal-title":"Expert Syst. Appl."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, D.V., and Rabinovich, A. (2015, January 7\u201312). Going deeper with convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"690","DOI":"10.1016\/j.ins.2015.10.002","article-title":"Feature selection with partition differentiation entropy for large-scale data sets","volume":"329","author":"Li","year":"2016","journal-title":"Inf. Sci."},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"162","DOI":"10.1109\/TAMD.2015.2431497","article-title":"Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks","volume":"7","author":"Zheng","year":"2015","journal-title":"IEEE Trans. Auton. Ment. Dev."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/14\/5158\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T23:47:25Z","timestamp":1760140045000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/14\/5158"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,7,9]]},"references-count":42,"journal-issue":{"issue":"14","published-online":{"date-parts":[[2022,7]]}},"alternative-id":["s22145158"],"URL":"https:\/\/doi.org\/10.3390\/s22145158","relation":{},"ISSN":["1424-8220"],"issn-type":[{"type":"electronic","value":"1424-8220"}],"subject":[],"published":{"date-parts":[[2022,7,9]]}}}