{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T04:40:17Z","timestamp":1743396017442,"version":"3.40.3"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2024,4,27]],"date-time":"2024-04-27T00:00:00Z","timestamp":1714176000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,27]],"date-time":"2024-04-27T00:00:00Z","timestamp":1714176000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-19201-z","type":"journal-article","created":{"date-parts":[[2024,4,27]],"date-time":"2024-04-27T04:02:20Z","timestamp":1714190540000},"page":"8037-8070","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Twin attention based multi-task convolutional bidirectional long short term memory for facial expression recognition"],"prefix":"10.1007","volume":"84","author":[{"given":"Velagapudi","family":"Sreenivas","sequence":"first","affiliation":[]},{"given":"B.","family":"Sivaneasan","sequence":"additional","affiliation":[]},{"given":"K. Suvarna","family":"Vani","sequence":"additional","affiliation":[]},{"given":"Prasun","family":"Chakrabarti","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,27]]},"reference":[{"key":"19201_CR1","doi-asserted-by":"crossref","first-page":"41273","DOI":"10.1109\/ACCESS.2019.2907327","volume":"7","author":"JH Kim","year":"2019","unstructured":"Kim JH, Kim BG, Roy PP, Jeong DM (2019) Efficient facial expression recognition algorithm based on hierarchical deep neural network structure. IEEE access 7:41273\u201341285","journal-title":"IEEE access"},{"issue":"9","key":"19201_CR2","doi-asserted-by":"crossref","first-page":"3046","DOI":"10.3390\/s21093046","volume":"21","author":"S Minaee","year":"2021","unstructured":"Minaee S, Minaei M, Abdolrashidi A (2021) Deep-emotion: facial expression recognition using attentional convolutional network. Sensors 21(9):3046","journal-title":"Sensors"},{"issue":"3","key":"19201_CR3","doi-asserted-by":"crossref","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li S, Deng W (2020) deep facial expression recognition: a survey. IEEE Trans Affect Comput 13(3):1195\u20131215","journal-title":"IEEE Trans Affect Comput"},{"issue":"1","key":"19201_CR4","doi-asserted-by":"crossref","first-page":"14","DOI":"10.18178\/ijmlc.2019.9.1.759","volume":"9","author":"WH Abdulsalam","year":"2019","unstructured":"Abdulsalam WH, Alhamdani RS, Abdullah MN (2019) Facial emotion recognition from videos using deep convolutional neural networks. Int J Mach Learn Comput 9(1):14\u201319","journal-title":"Int J Mach Learn Comput"},{"issue":"5","key":"19201_CR5","doi-asserted-by":"crossref","first-page":"1897","DOI":"10.3390\/app10051897","volume":"10","author":"Y Wang","year":"2020","unstructured":"Wang Y, Li Y, Song Y, Rong X (2020) The influence of the activation function in a convolution neural network model of facial expression recognition. Appl Sci 10(5):1897","journal-title":"Appl Sci"},{"key":"19201_CR6","doi-asserted-by":"crossref","unstructured":"Chen X, Yang X, Wang M, Zou J (2017) Convolution neural network for automatic facial expression recognition. In 2017 International conference on applied system innovation (ICASI) 814\u2013817. IEEE","DOI":"10.1109\/ICASI.2017.7988558"},{"issue":"3","key":"19201_CR7","doi-asserted-by":"crossref","first-page":"446","DOI":"10.1007\/s42452-020-2234-1","volume":"2","author":"N Mehendale","year":"2020","unstructured":"Mehendale N (2020) Facial emotion recognition using convolutional neural networks (FERC). SN Appl Sci 2(3):446","journal-title":"SN Appl Sci"},{"key":"19201_CR8","doi-asserted-by":"crossref","first-page":"26756","DOI":"10.1109\/ACCESS.2022.3156598","volume":"10","author":"AP Fard","year":"2022","unstructured":"Fard AP, Mahoor MH (2022) Ad-corre: adaptive correlation-based loss for facial expression recognition in the wild. IEEE Access 10:26756\u201326768","journal-title":"IEEE Access"},{"issue":"3","key":"19201_CR9","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1007\/s11554-023-01310-x","volume":"20","author":"CL Kim","year":"2023","unstructured":"Kim CL, Kim BG (2023) Few-shot learning for facial expression recognition: a comprehensive survey. J Real-Time Image Proc 20(3):52","journal-title":"J Real-Time Image Proc"},{"issue":"6","key":"19201_CR10","doi-asserted-by":"crossref","first-page":"4435","DOI":"10.1016\/j.aej.2021.09.066","volume":"61","author":"Y Nan","year":"2022","unstructured":"Nan Y, Ju J, Hua Q, Zhang H, Wang B (2022) A-MobileNet: an approach of facial expression recognition. Alex Eng J 61(6):4435\u20134444","journal-title":"Alex Eng J"},{"key":"19201_CR11","doi-asserted-by":"crossref","first-page":"103395","DOI":"10.1016\/j.jvcir.2021.103395","volume":"82","author":"W Dias","year":"2022","unstructured":"Dias W, Andalo F, Padilha R, Bertocco G, Almeida W, Costa P, Rocha A (2022) Cross-dataset emotion recognition from facial expressions through convolutional neural networks. J Vis Commun Image Represent 82:103395","journal-title":"J Vis Commun Image Represent"},{"key":"19201_CR12","doi-asserted-by":"crossref","unstructured":"Li H, Wang N, Yang X, Wang X, Gao X (2022) Towards semi-supervised deep facial expression recognition with an adaptive confidence margin. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition 4166\u20134175","DOI":"10.1109\/CVPR52688.2022.00413"},{"issue":"9","key":"19201_CR13","doi-asserted-by":"crossref","first-page":"7335","DOI":"10.1016\/j.jksuci.2021.08.021","volume":"34","author":"M Aslan","year":"2022","unstructured":"Aslan M (2022) CNN based efficient approach for emotion recognition. J King Saud Univ-Comput Inf Sci 34(9):7335\u20137346","journal-title":"J King Saud Univ-Comput Inf Sci"},{"issue":"01","key":"19201_CR14","doi-asserted-by":"crossref","first-page":"2250011","DOI":"10.1142\/S0218127422500110","volume":"32","author":"WD Dang","year":"2022","unstructured":"Dang WD, Lv DM, Li RM, Rui LG, Yang ZY, Ma C, Gao ZK (2022) Multilayer network-based CNN model for emotion recognition. Int J Bifurcation Chaos 32(01):2250011","journal-title":"Int J Bifurcation Chaos"},{"issue":"8","key":"19201_CR15","doi-asserted-by":"crossref","first-page":"2976","DOI":"10.3390\/s22082976","volume":"22","author":"M Algarni","year":"2022","unstructured":"Algarni M, Saeed F, Al-Hadhrami T, Ghabban F, Al-Sarem M (2022) Deep learning-based approach for emotion recognition using electroencephalography (EEG) signals using bi-directional long short-term memory (Bi-LSTM). Sensors 22(8):2976","journal-title":"Sensors"},{"issue":"2","key":"19201_CR16","doi-asserted-by":"crossref","first-page":"199","DOI":"10.3390\/biomimetics8020199","volume":"8","author":"Z Wen","year":"2023","unstructured":"Wen Z, Lin W, Wang T, Xu G (2023) Distract your attention: multi-head cross attention network for facial expression recognition. Biomimetics 8(2):199","journal-title":"Biomimetics"},{"issue":"1","key":"19201_CR17","first-page":"67","volume":"37","author":"H Li","year":"2023","unstructured":"Li H, Niu H, Zhu Z, Zhao F (2023) Intensity-aware loss for dynamic facial expression recognition in the wild. In Proc AAAI Conf Artif Intell 37(1):67\u201375","journal-title":"In Proc AAAI Conf Artif Intell"},{"issue":"9","key":"19201_CR18","doi-asserted-by":"crossref","first-page":"1036","DOI":"10.3390\/electronics10091036","volume":"10","author":"MA Akhand","year":"2021","unstructured":"Akhand MA, Roy S, Siddique N, Kamal MA, Shimamura T (2021) Facial emotion recognition using transfer learning in the deep CNN. Electronics 10(9):1036","journal-title":"Electronics"},{"key":"19201_CR19","first-page":"57","volume":"2","author":"B Li","year":"2021","unstructured":"Li B, Lima D (2021) Facial expression recognition via ResNet-50. Int J Cogn Comput Eng 2:57\u201364","journal-title":"Int J Cogn Comput Eng"},{"key":"19201_CR20","doi-asserted-by":"crossref","unstructured":"Echoukairi H, Ghmary ME, Ziani S, Ouacha A (2023) Improved methods for automatic facial expression recognition. Int J Interact Mob Technol 17(6)","DOI":"10.3991\/ijim.v17i06.37031"},{"key":"19201_CR21","doi-asserted-by":"crossref","first-page":"64827","DOI":"10.1109\/ACCESS.2019.2917266","volume":"7","author":"MI Georgescu","year":"2019","unstructured":"Georgescu MI, Ionescu RT, Popescu M (2019) Local learning with deep and handcrafted features for facial expression recognition. IEEE Access 7:64827\u201364836","journal-title":"IEEE Access"},{"key":"19201_CR22","unstructured":"Khaireddin Y, Chen Z (2021) Facial emotion recognition: State of the art performance on FER2013. arXivpreprint arXiv:2105.03588"},{"issue":"1","key":"19201_CR23","doi-asserted-by":"crossref","first-page":"502","DOI":"10.3390\/s23010502","volume":"23","author":"D Mamieva","year":"2023","unstructured":"Mamieva D, Abdusalomov AB, Mukhiddinov M, Whangbo TK (2023) Improved face detection method via learning small faces on hard images based on a deep learning approach. Sensors 23(1):502","journal-title":"Sensors"},{"issue":"3","key":"19201_CR24","doi-asserted-by":"crossref","first-page":"1080","DOI":"10.3390\/s23031080","volume":"23","author":"M Mukhiddinov","year":"2023","unstructured":"Mukhiddinov M, Djuraev O, Akhmedov F, Mukhamadiyev A, Cho J (2023) Masked face emotion recognition based on facial landmarks and deep learning approaches for visually impaired people. Sensors 23(3):1080","journal-title":"Sensors"},{"issue":"8","key":"19201_CR25","doi-asserted-by":"crossref","first-page":"11365","DOI":"10.1007\/s11042-022-13558-9","volume":"82","author":"S Gupta","year":"2023","unstructured":"Gupta S, Kumar P, Tekchandani RK (2023) Facial emotion recognition based real-time learner engagement detection system in online learning context using deep learning models. Multimed Tools Appl 82(8):11365\u201311394","journal-title":"Multimed Tools Appl"},{"key":"19201_CR26","doi-asserted-by":"crossref","first-page":"163677","DOI":"10.1016\/j.ijleo.2019.163677","volume":"208","author":"DN Thanh","year":"2020","unstructured":"Thanh DN, Hien NN, Prasath S (2020) Adaptive total variation L1 regularization for salt and pepper image denoising. Optik 208:163677","journal-title":"Optik"},{"key":"19201_CR27","doi-asserted-by":"crossref","unstructured":"Simarmata TS, Isnanto RR, Triwiyatno A (2023) Detection of pulmonary tuberculosis using neural network with feature extraction of gray level run-length matrix method on lung x-ray images. In: 2023 International Seminar on Intelligent Technology and Its Applications (ISITIA). IEEE, pp 33\u201344","DOI":"10.1109\/ISITIA59021.2023.10221153"},{"key":"19201_CR28","doi-asserted-by":"crossref","first-page":"103079","DOI":"10.1016\/j.cviu.2020.103079","volume":"201","author":"Q Yan","year":"2020","unstructured":"Yan Q, Wang B, Li P, Li X, Zhang A, Shi Q, You Z, Zhu Y, Sun J, Zhang Y (2020) Ghost removal via channel attention in exposure fusion. Comput Vis Image Underst 201:103079","journal-title":"Comput Vis Image Underst"},{"key":"19201_CR29","first-page":"1","volume":"19","author":"X Yang","year":"2021","unstructured":"Yang X, Zhang M, Li W, Tao R (2021) Visible-assisted infrared image super-resolution based on spatial attention residual network. IEEE Geosci Remote Sens Lett 19:1\u20135","journal-title":"IEEE Geosci Remote Sens Lett"},{"key":"19201_CR30","doi-asserted-by":"crossref","first-page":"114194","DOI":"10.1016\/j.cma.2021.114194","volume":"388","author":"W Zhao","year":"2022","unstructured":"Zhao W, Wang L, Mirjalili S (2022) artificial hummingbird algorithm: a new bio-inspired optimizer with its engineering applications. Comput Methods Appl Mech Eng 388:114194","journal-title":"Comput Methods Appl Mech Eng"},{"key":"19201_CR31","doi-asserted-by":"crossref","first-page":"1075","DOI":"10.1007\/s11036-020-01530-6","volume":"25","author":"M Roccetti","year":"2020","unstructured":"Roccetti M, Delnevo G, Casini L, Salomoni P (2020) a cautionary tale for machine learning design: why we still need human-assisted big data analysis. Mob Netw Appl 25:1075\u20131083","journal-title":"Mob Netw Appl"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19201-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-19201-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19201-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T04:18:14Z","timestamp":1743394694000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-19201-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,27]]},"references-count":31,"journal-issue":{"issue":"10","published-online":{"date-parts":[[2025,3]]}},"alternative-id":["19201"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-19201-z","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,4,27]]},"assertion":[{"value":"13 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 April 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 April 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 April 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Authors declare that they have no conflict of interest.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All the authors involved have agreed to participate in this submitted article.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"All the authors involved in this manuscript give full consent for publication of this submitted article.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to publish"}}]}}