{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,14]],"date-time":"2026-01-14T19:47:43Z","timestamp":1768420063397,"version":"3.49.0"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T00:00:00Z","timestamp":1734652800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T00:00:00Z","timestamp":1734652800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-024-03537-2","type":"journal-article","created":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T12:18:10Z","timestamp":1734697090000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Lightweight Attention Based Deep CNN Framework for Human Facial Emotion Detection from Video Sequences"],"prefix":"10.1007","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3868-1895","authenticated-orcid":false,"given":"Krishna","family":"Kant","sequence":"first","affiliation":[]},{"given":"Dipti B.","family":"Shah","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,20]]},"reference":[{"issue":"9","key":"3537_CR1","doi-asserted-by":"publisher","first-page":"1036","DOI":"10.3390\/electronics10091036","volume":"10","author":"M Akhand","year":"2021","unstructured":"Akhand M, Roy S, Siddique N. Facial emotion recognition using transfer learning in the deep cnn. Electronics. 2021;10(9):1036.","journal-title":"Electronics"},{"key":"3537_CR2","doi-asserted-by":"publisher","first-page":"979","DOI":"10.1016\/j.procs.2020.03.397","volume":"167","author":"I Ali","year":"2020","unstructured":"Ali I, Dua M. Smile detection using data amalgamation. Procedia Comput Sci. 2020;167:979\u201386.","journal-title":"Procedia Comput Sci"},{"key":"3537_CR3","doi-asserted-by":"publisher","first-page":"110494","DOI":"10.1016\/j.asoc.2023.110494","volume":"144","author":"A Aslam","year":"2023","unstructured":"Aslam A, Sargano AB, Habib Z. Attention-based multimodal sentiment analysis and emotion recognition using deep neural networks. Appl Soft Comput. 2023;144:110494.","journal-title":"Appl Soft Comput"},{"issue":"1","key":"3537_CR4","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1177\/1754073918768878","volume":"11","author":"NH Bailen","year":"2019","unstructured":"Bailen NH, Green LM, Thompson RJ. Understanding emotion in adolescents: a review of emotional frequency, intensity, instability, and clarity. Emot Rev. 2019;11(1):63\u201373.","journal-title":"Emot Rev"},{"issue":"8","key":"3537_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10462-024-10831-1","volume":"57","author":"T Chutia","year":"2024","unstructured":"Chutia T, Baruah N. A review on emotion detection by using deep learning techniques. Artif Intell Rev. 2024;57(8):1\u201380.","journal-title":"Artif Intell Rev"},{"issue":"01","key":"3537_CR6","doi-asserted-by":"publisher","first-page":"1550006","DOI":"10.1142\/S0218001415500068","volume":"29","author":"D Freire-Obregon","year":"2015","unstructured":"Freire-Obregon D, Castrillon-Santana M. An evolutive approach for smile recog-\u00b4 nition in video sequences. Int J Pattern Recognit Artif Intell. 2015;29(01):1550006.","journal-title":"Int J Pattern Recognit Artif Intell"},{"key":"3537_CR7","doi-asserted-by":"publisher","first-page":"102218","DOI":"10.1016\/j.inffus.2023.102218","volume":"105","author":"A Geetha","year":"2024","unstructured":"Geetha A, Mala T, Priyanka D, et al. Multimodal emotion recognition with deep learning: advancements, challenges, and future directions. Inf Fusion. 2024;105:102218.","journal-title":"Inf Fusion"},{"key":"3537_CR8","doi-asserted-by":"publisher","DOI":"10.1007\/s13369-023-08538-6","author":"W Gong","year":"2024","unstructured":"Gong W, La Z, Qian Y, et al. Hybrid attention-aware learning network for facial expression recognition in the wild. Arab J Sci Eng. 2024. https:\/\/doi.org\/10.1007\/s13369-023-08538-6.","journal-title":"Arab J Sci Eng"},{"issue":"5","key":"3537_CR9","doi-asserted-by":"publisher","first-page":"1049","DOI":"10.1007\/s11760-020-01830-0","volume":"15","author":"N Hajarolasvadi","year":"2021","unstructured":"Hajarolasvadi N, Bashirov E, Demirel H. Video-based person-dependent and person-independent facial emotion recognition. SIViP. 2021;15(5):1049\u201356.","journal-title":"SIViP"},{"key":"3537_CR10","doi-asserted-by":"publisher","first-page":"108339","DOI":"10.1016\/j.engappai.2024.108339","volume":"133","author":"S Hazmoune","year":"2024","unstructured":"Hazmoune S, Bougamouza F. Using transformers for multimodal emotion recognition: taxonomies and state of the art review. Eng Appl Artif Intell. 2024;133:108339.","journal-title":"Eng Appl Artif Intell"},{"key":"3537_CR11","doi-asserted-by":"publisher","first-page":"165","DOI":"10.1016\/j.neucom.2020.10.015","volume":"422","author":"L He","year":"2021","unstructured":"He L, Chan JCW, Wang Z. Automatic depression recognition using cnn with attention mechanism from videos. Neurocomputing. 2021;422:165\u201375.","journal-title":"Neurocomputing"},{"issue":"2","key":"3537_CR12","doi-asserted-by":"publisher","first-page":"280","DOI":"10.1037\/0022-3514.69.2.280","volume":"69","author":"U Hess","year":"1995","unstructured":"Hess U, Banse R, Kappas A. The intensity of facial expression is determined by underlying affective state and social situation. J Pers Soc Psychol. 1995;69(2):280.","journal-title":"J Pers Soc Psychol"},{"key":"3537_CR13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243","volume-title":"Densely connected convolutional networks","author":"G Huang","year":"2017","unstructured":"Huang G, Liu Z, Van Der Maaten L, et al. Densely connected convolutional networks. New York: IEEE; 2017."},{"key":"3537_CR14","doi-asserted-by":"publisher","first-page":"1132","DOI":"10.1016\/j.neucom.2017.09.056","volume":"275","author":"X Jiang","year":"2018","unstructured":"Jiang X, Pang Y, Li X, et al. Deep neural networks with elastic rectified linear units for object recognition. Neurocomputing. 2018;275:1132\u20139.","journal-title":"Neurocomputing"},{"key":"3537_CR15","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1016\/j.patrec.2017.04.003","volume":"92","author":"SKA Kamarol","year":"2017","unstructured":"Kamarol SKA, Jaward MH, Kalviainen H, et al. Joint facial expression recog-\u00a8 nition and intensity estimation based on weighted votes of image sequences. Pattern Recognit Lett. 2017;92:25\u201332.","journal-title":"Pattern Recognit Lett"},{"issue":"3","key":"3537_CR16","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1007\/s00530-024-01302-2","volume":"30","author":"UA Khan","year":"2024","unstructured":"Khan UA, Xu Q, Liu Y, et al. Exploring contactless techniques in multimodal emotion recognition: insights into diverse applications, challenges, solutions, and prospects. Multimed Syst. 2024;30(3):115.","journal-title":"Multimed Syst"},{"key":"3537_CR17","doi-asserted-by":"publisher","first-page":"340","DOI":"10.1016\/j.neucom.2020.06.014","volume":"411","author":"J Li","year":"2020","unstructured":"Li J, Jin K, Zhou D, et al. Attention mechanism-based cnn for facial expression recognition. Neurocomputing. 2020;411:340\u201350.","journal-title":"Neurocomputing"},{"key":"3537_CR18","first-page":"8843413","volume":"1","author":"X Liu","year":"2020","unstructured":"Liu X. Wang M (2020) Context-aware attention network for human emotion recognition in video. Adv Multimed. 2020;1:8843413.","journal-title":"Adv Multimed"},{"key":"3537_CR19","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1016\/j.ins.2022.03.062","volume":"598","author":"Y Liu","year":"2022","unstructured":"Liu Y, Feng C, Yuan X, et al. Clip-aware expressive feature learning for videobased facial expression recognition. Inf Sci. 2022;598:182\u201395.","journal-title":"Inf Sci"},{"issue":"2","key":"3537_CR20","doi-asserted-by":"publisher","first-page":"4667","DOI":"10.1007\/s11042-023-15403-z","volume":"83","author":"G Lu","year":"2024","unstructured":"Lu G, Chen H, Wei J, et al. Video-based neonatal pain expression recognition with cross-stream attention. Multimed Tools Appl. 2024;83(2):4667\u201390.","journal-title":"Multimed Tools Appl"},{"issue":"3","key":"3537_CR21","doi-asserted-by":"publisher","first-page":"8207","DOI":"10.1007\/s11042-023-15762-7","volume":"83","author":"Y Luo","year":"2024","unstructured":"Luo Y, Wu R, Liu J, et al. Attention fusion network for multimodal sentiment analysis. Multimed Tools Appl. 2024;83(3):8207\u201317.","journal-title":"Multimed Tools Appl"},{"issue":"3","key":"3537_CR22","doi-asserted-by":"publisher","first-page":"399","DOI":"10.1093\/oxfordjournals.schbul.a033335","volume":"24","author":"MK Mandal","year":"1998","unstructured":"Mandal MK, Pandey R, Prasad AB. Facial expressions of emotions and schizophrenia: a review. Schizophr Bull. 1998;24(3):399\u2013412.","journal-title":"Schizophr Bull"},{"issue":"3","key":"3537_CR23","doi-asserted-by":"publisher","first-page":"399","DOI":"10.2979\/VIC.2008.50.3.399","volume":"50","author":"J Mayer","year":"2008","unstructured":"Mayer J. The expression of the emotions in man and laboratory animals. Vic Stud. 2008;50(3):399\u2013417.","journal-title":"Vic Stud"},{"issue":"6","key":"3537_CR24","doi-asserted-by":"publisher","first-page":"15711","DOI":"10.1007\/s11042-023-16174-3","volume":"83","author":"G Meena","year":"2024","unstructured":"Meena G, Mohbey KK, Indian A, et al. Identifying emotions from facial expressions using a deep convolutional neural network-based approach. Multimed Tools Appl. 2024;83(6):15711\u201332.","journal-title":"Multimed Tools Appl"},{"key":"3537_CR25","volume-title":"Nonverbal communication, book-non-verbal communication","author":"A Mehrabian","year":"2017","unstructured":"Mehrabian A. Nonverbal communication, book-non-verbal communication. London: Routledge; 2017."},{"issue":"8","key":"3537_CR26","doi-asserted-by":"publisher","first-page":"1897","DOI":"10.3390\/s19081897","volume":"19","author":"D Mehta","year":"2019","unstructured":"Mehta D, Siddiqui MFH, Javaid AY. Recognition of emotion intensities using machine learning algorithms: a comparative study. Sensors. 2019;19(8):1897.","journal-title":"Sensors"},{"key":"3537_CR27","doi-asserted-by":"publisher","first-page":"114693","DOI":"10.1016\/j.eswa.2021.114693","volume":"173","author":"L Mou","year":"2021","unstructured":"Mou L, Zhou C, Zhao P, et al. Driver stress detection via multimodal fusion using attention-based cnn-lstm. Expert Syst Appl. 2021;173:114693.","journal-title":"Expert Syst Appl"},{"key":"3537_CR28","doi-asserted-by":"crossref","unstructured":"Mukherjee S, Vamshi B, Reddy KSVK et al (2016) Recognizing facial expressions using novel motion based features. pp. 1\u20138","DOI":"10.1145\/3009977.3010004"},{"key":"3537_CR29","doi-asserted-by":"publisher","first-page":"122946","DOI":"10.1016\/j.eswa.2023.122946","volume":"245","author":"K Mustaqeem","year":"2024","unstructured":"Mustaqeem K, Gueaieb W, El Saddik A, et al. Mser: Multimodal speech emotion recognition using cross-attention with deep fusion. Expert Syst Appl. 2024;245:122946.","journal-title":"Expert Syst Appl"},{"key":"3537_CR30","first-page":"1949","volume":"5","author":"R Pal","year":"2016","unstructured":"Pal R, Satsangi C. Facial expression recognition based on basic expressions and intensities using k-means clustering. Int J Sci Res. 2016;5:1949\u201352.","journal-title":"Int J Sci Res"},{"key":"3537_CR31","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1016\/j.cogsys.2020.03.002","volume":"62","author":"GV Reddy","year":"2020","unstructured":"Reddy GV, Savarni CD, Mukherjee S. Facial expression recognition in the wild, by fusion of deep learnt and hand-crafted features. Cogn Syst Res. 2020;62:23\u201334.","journal-title":"Cogn Syst Res"},{"key":"3537_CR32","doi-asserted-by":"publisher","first-page":"103370","DOI":"10.1016\/j.advengsoft.2022.103370","volume":"176","author":"P Singh","year":"2023","unstructured":"Singh P, Muchahari MK. Solving multi-objective optimization problem of convolutional neural network using fast forward quantum optimization algorithm: application in digital image classification. Adv Eng Softw. 2023;176:103370.","journal-title":"Adv Eng Softw"},{"issue":"4","key":"3537_CR33","first-page":"1819","volume":"15","author":"R Singh","year":"2023","unstructured":"Singh R, Saurav S, Kumar T, et al. Facial expression recognition in videos using hybrid cnn & convlstm. Int J Inf Technol. 2023;15(4):1819\u201330.","journal-title":"Int J Inf Technol"},{"issue":"2","key":"3537_CR34","doi-asserted-by":"publisher","first-page":"543","DOI":"10.1007\/s10044-022-01124-w","volume":"26","author":"C Su","year":"2023","unstructured":"Su C, Wei J, Lin D, et al. Using attention lsgb network for facial expression recognition. Pattern Anal Appl. 2023;26(2):543\u201353.","journal-title":"Pattern Anal Appl"},{"key":"3537_CR35","doi-asserted-by":"publisher","DOI":"10.1007\/s13042-024-02287-0","author":"H Tang","year":"2024","unstructured":"Tang H, Li Y, Jin Z. A dual stream attention network for facial expression recognition in the wild. Int J Mach Learn Cyber. 2024. https:\/\/doi.org\/10.1007\/s13042-024-02287-0.","journal-title":"Int J Mach Learn Cyber"},{"key":"3537_CR36","doi-asserted-by":"publisher","first-page":"123834","DOI":"10.1016\/j.eswa.2024.123834","volume":"249","author":"JP Thekkekara","year":"2024","unstructured":"Thekkekara JP, Yongchareon S, Liesaputra V. An attention-based cnn-bilstm model for depression detection on social media text. Expert Syst Appl. 2024;249:123834.","journal-title":"Expert Syst Appl"},{"key":"3537_CR37","first-page":"387","volume-title":"Detecting micro-expression intensity changes from videos based on hybrid deep cnn","author":"S Thuseethan","year":"2019","unstructured":"Thuseethan S, Rajasegarar S, Yearwood J. Detecting micro-expression intensity changes from videos based on hybrid deep cnn. Cham: Springer; 2019. p. 387\u201399."},{"key":"3537_CR38","first-page":"1","volume-title":"Emotion intensity estimation from video frames using deep hybrid convolutional neural networks","author":"S Thuseethan","year":"2019","unstructured":"Thuseethan S, Rajasegarar S, Yearwood J. Emotion intensity estimation from video frames using deep hybrid convolutional neural networks. New York: IEEE; 2019. p. 1\u201310."},{"key":"3537_CR39","doi-asserted-by":"publisher","first-page":"370","DOI":"10.1016\/j.ins.2023.02.056","volume":"630","author":"Z Zhang","year":"2023","unstructured":"Zhang Z, Tian X, Zhang Y, et al. Enhanced discriminative global-local feature learning with priority for facial expression recognition. Inf Sci. 2023;630:370\u201384.","journal-title":"Inf Sci"},{"issue":"3","key":"3537_CR40","doi-asserted-by":"publisher","first-page":"224","DOI":"10.1109\/TETCI.2017.2778716","volume":"2","author":"K Zheng","year":"2017","unstructured":"Zheng K, Yan WQ, Nand P. Video dynamics detection using deep neural networks. IEEE Trans Emerg Topics Computat Intell. 2017;2(3):224\u201334.","journal-title":"IEEE Trans Emerg Topics Computat Intell"}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-024-03537-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-024-03537-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-024-03537-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T13:02:49Z","timestamp":1734699769000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-024-03537-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,20]]},"references-count":40,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,1]]}},"alternative-id":["3537"],"URL":"https:\/\/doi.org\/10.1007\/s42979-024-03537-2","relation":{},"ISSN":["2661-8907"],"issn-type":[{"value":"2661-8907","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,20]]},"assertion":[{"value":"23 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that none of the work reported in this study could have been influenced by any known competing financial interests or personal relationships. The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"I declare that I have not used any images or tables or data from other\u2019s research work. This work has been solely contributed by me and my guide(Dr.Dipti.B.Shah, Professor, PG Department of Computer Science and Technology , Sardar Patel University, Anand).","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed Consent"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Research Involving Human and\/or Animals"}}],"article-number":"22"}}