{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T14:20:33Z","timestamp":1775744433738,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,3,29]],"date-time":"2026-03-29T00:00:00Z","timestamp":1774742400000},"content-version":"vor","delay-in-days":28,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s00371-026-04396-z","type":"journal-article","created":{"date-parts":[[2026,3,29]],"date-time":"2026-03-29T08:27:43Z","timestamp":1774772863000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Attention-enhanced CNN-LSTM framework for real-time video-based emotion recognition"],"prefix":"10.1007","volume":"42","author":[{"given":"Anandhavalli","family":"Muniasamy","sequence":"first","affiliation":[]},{"given":"Rizwan","family":"Abbas","sequence":"additional","affiliation":[]},{"given":"Galiya","family":"Ybytayeva","sequence":"additional","affiliation":[]},{"given":"Ashwag","family":"Alasmari","sequence":"additional","affiliation":[]},{"given":"Nouf","family":"Aldahwan","sequence":"additional","affiliation":[]},{"given":"Hend Khalid","family":"Alkahtani","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,29]]},"reference":[{"key":"4396_CR1","doi-asserted-by":"publisher","unstructured":"Paiva, A.: Robots that listen to people\u2019s hearts: the role of emotions in the communication between humans and social robots. In: Proceedings of the 26th Conference on User Modeling, Adaptation and Personalization. UMAP \u201918, p. 175. Association for Computing Machinery, New York, NY, USA (2018). https:\/\/doi.org\/10.1145\/3209219.3209268","DOI":"10.1145\/3209219.3209268"},{"issue":"4","key":"4396_CR2","doi-asserted-by":"publisher","first-page":"3040","DOI":"10.1007\/S10489-024-05329-W","volume":"54","author":"C Shi","year":"2024","unstructured":"Shi, C., Zhang, Y., Liu, B.: A multimodal fusion-based deep learning framework combined with local-global contextual TCNs for continuous emotion recognition from videos. Appl. Intell. 54(4), 3040\u20133057 (2024). https:\/\/doi.org\/10.1007\/S10489-024-05329-W","journal-title":"Appl. Intell."},{"key":"4396_CR3","doi-asserted-by":"publisher","first-page":"106608","DOI":"10.1016\/J.BSPC.2024.106608","volume":"96","author":"X Tao","year":"2024","unstructured":"Tao, X., Su, L., Rao, Z., Li, Y., Wu, D., Ji, X., Liu, J.: Facial video-based non-contact emotion recognition: a multi-view features expression and fusion method. Biomed. Signal Process. Control 96, 106608 (2024). https:\/\/doi.org\/10.1016\/J.BSPC.2024.106608","journal-title":"Biomed. Signal Process. Control"},{"key":"4396_CR4","doi-asserted-by":"publisher","first-page":"105062","DOI":"10.1016\/J.COMPEDU.2024.105062","volume":"218","author":"C Zhang","year":"2024","unstructured":"Zhang, C., Wang, Z., Fang, Z., Xiao, X.: Guiding student learning in video lectures: effects of instructors\u2019 emotional expressions and visual cues. Comput. Educ. 218, 105062 (2024). https:\/\/doi.org\/10.1016\/J.COMPEDU.2024.105062","journal-title":"Comput. Educ."},{"key":"4396_CR5","doi-asserted-by":"publisher","unstructured":"Wei, J., Hu, G., Yang, X., Luu, A.T., Dong, Y.: Learning facial expression and body gesture visual information for video emotion recognition. Expert Syst. Appl. 237(Part A), 121419 (2024). https:\/\/doi.org\/10.1016\/J.ESWA.2023.121419","DOI":"10.1016\/J.ESWA.2023.121419"},{"key":"4396_CR6","doi-asserted-by":"publisher","unstructured":"Sun, Z., Xuan, Y., Liu, F., Xiang, Y.: FG-EmoTalk: talking head video generation with fine-grained controllable facial expressions. Proc. AAAI Conf. Artif. Intell. 38(5), 5043\u20135051 (2024). https:\/\/doi.org\/10.1609\/aaai.v38i5.28309","DOI":"10.1609\/aaai.v38i5.28309"},{"key":"4396_CR7","doi-asserted-by":"publisher","unstructured":"Bissinger, B., Beer, A., M\u00e4rtin, C., Fellmann, M.: Emotion recognition via facial expressions to improve virtual communication in videoconferences. In: Human-Computer Interaction: Thematic Area, HCI 2023, Held as Part of the 25th HCI International Conference, HCII 2023, Copenhagen, Denmark, July 23\u201328, 2023, Proceedings, Part II, pp. 151\u2013163. Springer, Berlin (2023). https:\/\/doi.org\/10.1007\/978-3-031-35599-8_10","DOI":"10.1007\/978-3-031-35599-8_10"},{"key":"4396_CR8","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/J.NEUNET.2014.09.005","volume":"64","author":"IJ Goodfellow","year":"2015","unstructured":"...Goodfellow, I.J., Erhan, D., Carrier, P.L., Courville, A.C., Mirza, M., Hamner, B., Cukierski, W., Tang, Y., Thaler, D., Lee, D., Zhou, Y., Ramaiah, C., Feng, F., Li, R., Wang, X., Athanasakis, D., Shawe-Taylor, J., Milakov, M., Park, J., Ionescu, R.T., Popescu, M., Grozea, C., Bergstra, J., Xie, J., Romaszko, L., Xu, B., Chuang, Z., Bengio, Y.: Challenges in representation learning: a report on three machine learning contests. Neural Netw. 64, 59\u201363 (2015). https:\/\/doi.org\/10.1016\/J.NEUNET.2014.09.005","journal-title":"Neural Netw."},{"issue":"3","key":"4396_CR9","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1109\/MMUL.2012.26","volume":"19","author":"A Dhall","year":"2012","unstructured":"Dhall, A., Goecke, R., Lucey, S., Gedeon, T.: Collecting large, richly annotated facial-expression databases from movies. IEEE Multim. 19(3), 34\u201341 (2012). https:\/\/doi.org\/10.1109\/MMUL.2012.26","journal-title":"IEEE Multim."},{"issue":"1","key":"4396_CR10","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/TIP.2018.2868382","volume":"28","author":"S Li","year":"2019","unstructured":"Li, S., Deng, W.: Reliable crowdsourcing and deep locality-preserving learning for unconstrained facial expression recognition. IEEE Trans. Image Process. 28(1), 356\u2013370 (2019). https:\/\/doi.org\/10.1109\/TIP.2018.2868382","journal-title":"IEEE Trans. Image Process."},{"key":"4396_CR11","doi-asserted-by":"publisher","unstructured":"Farman, H., Sedik, A., Nasralla, M.M., Esmail, M.A.: Facial emotion recognition in smart education systems: a review. In: 2023 IEEE International Smart Cities Conference (ISC2), pp. 1\u20139 (2023). https:\/\/doi.org\/10.1109\/ISC257844.2023.10293353","DOI":"10.1109\/ISC257844.2023.10293353"},{"key":"4396_CR12","doi-asserted-by":"publisher","first-page":"100798","DOI":"10.1016\/J.ENTCOM.2024.100798","volume":"52","author":"S Li","year":"2025","unstructured":"Li, S.: Application of entertainment e-learning mode based on genetic algorithm and facial emotion recognition in environmental art and design courses. Entertain. Comput. 52, 100798 (2025). https:\/\/doi.org\/10.1016\/J.ENTCOM.2024.100798","journal-title":"Entertain. Comput."},{"key":"4396_CR13","doi-asserted-by":"publisher","unstructured":"Hadjar, H., Reis, T., Bornschlegl, M.X., Engel, F.C., Mc\u00a0Kevitt, P., Hemmje, M.L.: Recognition and visualization of facial expression and emotion in healthcare. In: Advanced Visual Interfaces. Supporting Artificial Intelligence and Big Data Applications: AVI 2020 Workshops, AVI-BDA and ITAVIS, Ischia, Italy, June 9, 2020 and September 29, 2020, Revised Selected Papers, pp. 109\u2013124. Springer, Berlin, Heidelberg (2020). https:\/\/doi.org\/10.1007\/978-3-030-68007-7_7","DOI":"10.1007\/978-3-030-68007-7_7"},{"key":"4396_CR14","doi-asserted-by":"publisher","unstructured":"Tzafilkou, K., Panavou, F.R., Economides, A.A.: Facially expressed emotions and hedonic liking on social media food marketing campaigns:comparing different types of products and media posts. In: 2022 17th International Workshop on Semantic and Social Media Adaptation & Personalization (SMAP), pp. 1\u20136 (2022). https:\/\/doi.org\/10.1109\/SMAP56125.2022.9942096","DOI":"10.1109\/SMAP56125.2022.9942096"},{"key":"4396_CR15","doi-asserted-by":"publisher","unstructured":"Wu, Y., Arevalillo\u00a0Herr\u00e1ez, M., Katsigiannis, S., Ramzan, N.: On the benefits of using hidden Markov models to predict emotions. In: Proceedings of the 30th ACM Conference on User Modeling, Adaptation and Personalization. UMAP \u201922, pp. 164\u2013169. Association for Computing Machinery, New York, NY, USA (2022). https:\/\/doi.org\/10.1145\/3503252.3531323","DOI":"10.1145\/3503252.3531323"},{"key":"4396_CR16","doi-asserted-by":"crossref","unstructured":"Postawka, A.: Behavior-based emotion recognition using Kinect and hidden Markov models. In: Rutkowski, L., Scherer, R., Korytkowski, M., Pedrycz, W., Tadeusiewicz, R., Zurada, J.M. (eds.) Artificial Intelligence and Soft Computing, pp. 250\u2013259. Springer, Cham (2019)","DOI":"10.1007\/978-3-030-20915-5_23"},{"key":"4396_CR17","doi-asserted-by":"publisher","unstructured":"Iriya, R., Ram\u00edrez, M.A.: Gaussian mixture models with class-dependent features for speech emotion recognition. In: 2014 IEEE Workshop on Statistical Signal Processing (SSP), pp. 480\u2013483 (2014). https:\/\/doi.org\/10.1109\/SSP.2014.6884680","DOI":"10.1109\/SSP.2014.6884680"},{"key":"4396_CR18","doi-asserted-by":"publisher","unstructured":"Tashev, I.J., Wang, Z.-Q., Godin, K.: Speech emotion recognition based on gaussian mixture models and deep neural networks. In: 2017 Information Theory and Applications Workshop (ITA), pp. 1\u20134 (2017). https:\/\/doi.org\/10.1109\/ITA.2017.8023477","DOI":"10.1109\/ITA.2017.8023477"},{"issue":"12","key":"4396_CR19","doi-asserted-by":"publisher","first-page":"4150","DOI":"10.1007\/S10489-019-01500-W","volume":"49","author":"MP Kumar","year":"2019","unstructured":"Kumar, M.P., Rajagopal, M.K.: Detecting facial emotions using normalized minimal feature vectors and semi-supervised twin support vector machines classifier. Appl. Intell. 49(12), 4150\u20134174 (2019). https:\/\/doi.org\/10.1007\/S10489-019-01500-W","journal-title":"Appl. Intell."},{"key":"4396_CR20","doi-asserted-by":"publisher","first-page":"104886","DOI":"10.1016\/J.KNOSYS.2019.104886","volume":"184","author":"A Bhavan","year":"2019","unstructured":"Bhavan, A., Chauhan, P., Shah, R.R.: Bagged support vector machines for emotion recognition from speech. Knowl. Based Syst. 184, 104886 (2019). https:\/\/doi.org\/10.1016\/J.KNOSYS.2019.104886","journal-title":"Knowl. Based Syst."},{"key":"4396_CR21","doi-asserted-by":"publisher","first-page":"139988","DOI":"10.1109\/ACCESS.2024.3450633","volume":"12","author":"LE Arenas-Deseano","year":"2024","unstructured":"Arenas-Deseano, L.E., Ram\u00edrez-Cort\u00e9s, J.M., de Jesus Rangel-Magdaleno, J., Cruz-Vega, I.: Real-time multiplatform emotion classification using CNN in a fog computing environment. IEEE Access 12, 139988\u2013139997 (2024). https:\/\/doi.org\/10.1109\/ACCESS.2024.3450633","journal-title":"IEEE Access"},{"key":"4396_CR22","doi-asserted-by":"publisher","first-page":"103268","DOI":"10.1016\/J.INFFUS.2025.103268","volume":"123","author":"X Zhu","year":"2025","unstructured":"Zhu, X., Wang, Y., Cambria, E., Rida, I., L\u00f3pez, J.S., Cui, L., Wang, R.: RMER-DT: robust multimodal emotion recognition in conversational contexts based on diffusion and transformers. Inf. Fusion 123, 103268 (2025). https:\/\/doi.org\/10.1016\/J.INFFUS.2025.103268","journal-title":"Inf. Fusion"},{"key":"4396_CR23","doi-asserted-by":"publisher","first-page":"100445","DOI":"10.1016\/J.ARRAY.2025.100445","volume":"27","author":"R Wang","year":"2025","unstructured":"Wang, R., Xu, D., Cascone, L., Wang, Y., Chen, H., Zheng, J., Zhu, X.: RAFT: robust adversarial fusion transformer for multimodal sentiment analysis. Array 27, 100445 (2025). https:\/\/doi.org\/10.1016\/J.ARRAY.2025.100445","journal-title":"Array"},{"key":"4396_CR24","doi-asserted-by":"publisher","first-page":"102860","DOI":"10.1109\/ACCESS.2022.3209813","volume":"10","author":"C-T Yen","year":"2022","unstructured":"Yen, C.-T., Li, K.-H.: Discussions of different deep transfer learning models for emotion recognitions. IEEE Access 10, 102860\u2013102875 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2022.3209813","journal-title":"IEEE Access"},{"key":"4396_CR25","doi-asserted-by":"publisher","unstructured":"Chandra, J., B., A., R., R.A.: Cross-database facial expression recognition using CNN with attention mechanism. In: 2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), pp. 1\u20137 (2023). https:\/\/doi.org\/10.1109\/ICCCNT56998.2023.10308238","DOI":"10.1109\/ICCCNT56998.2023.10308238"},{"issue":"2","key":"4396_CR26","doi-asserted-by":"publisher","first-page":"641","DOI":"10.1007\/S12530-023-09506-Z","volume":"15","author":"B Bakariya","year":"2024","unstructured":"Bakariya, B., Singh, A., Singh, H., Raju, P., Rajpoot, R., Mohbey, K.K.: Facial emotion recognition and music recommendation system using CNN-based deep learning techniques. Evol. Syst. 15(2), 641\u2013658 (2024). https:\/\/doi.org\/10.1007\/S12530-023-09506-Z","journal-title":"Evol. Syst."},{"issue":"4","key":"4396_CR27","doi-asserted-by":"publisher","first-page":"1819","DOI":"10.1007\/s41870-023-01183-0","volume":"15","author":"R Singh","year":"2023","unstructured":"Singh, R., Saurav, S., Kumar, T., Saini, R., Vohra, A., Singh, S.: Facial expression recognition in videos using hybrid CNN & ConvLSTM. Int. J. Inf. Technol. 15(4), 1819\u20131830 (2023). https:\/\/doi.org\/10.1007\/s41870-023-01183-0","journal-title":"Int. J. Inf. Technol."},{"key":"4396_CR28","doi-asserted-by":"crossref","unstructured":"Lamba, P.S., Virmani, D.: Cnn-lstm-based facial expression recognition. In: Abraham, A., Castillo, O., Virmani, D. (eds.) Proceedings of 3rd International Conference on Computing Informatics and Networks, pp. 379\u2013389. Springer, Singapore (2021)","DOI":"10.1007\/978-981-15-9712-1_32"},{"key":"4396_CR29","doi-asserted-by":"publisher","first-page":"549","DOI":"10.1007\/978-981-15-7527-3_52","volume-title":"Research in Intelligent and Computing in Engineering","author":"BT Hung","year":"2021","unstructured":"Hung, B.T., Tien, L.M.: Facial expression recognition with CNN-LSTM. In: Kumar, R., Quang, N.H., Kumar Solanki, V., Cardona, M., Pattnaik, P.K. (eds.) Research in Intelligent and Computing in Engineering, pp. 549\u2013560. Springer, Singapore (2021)"},{"key":"4396_CR30","doi-asserted-by":"publisher","unstructured":"G\u00f3mez-Sirvent, J.L., L\u00f3pez de\u00a0la Rosa, F., L\u00f3pez, M.T., Fern\u00e1ndez-Caballero, A.: Facial expression recognition in the wild for low-resolution images using voting residual network. Electronics 12(18), 3837 (2023). https:\/\/doi.org\/10.3390\/electronics12183837","DOI":"10.3390\/electronics12183837"},{"key":"4396_CR31","doi-asserted-by":"publisher","first-page":"45543","DOI":"10.1109\/ACCESS.2024.3380847","volume":"12","author":"MC Gursesli","year":"2024","unstructured":"Gursesli, M.C., Lombardi, S., Duradoni, M., Bocchi, L., Guazzini, A., Lanata, A.: Facial emotion recognition (fer) through custom lightweight CNN model: performance evaluation in public datasets. IEEE Access 12, 45543\u201345559 (2024). https:\/\/doi.org\/10.1109\/ACCESS.2024.3380847","journal-title":"IEEE Access"},{"issue":"3","key":"4396_CR32","doi-asserted-by":"publisher","first-page":"446","DOI":"10.1007\/s42452-020-2234-1","volume":"2","author":"N Mehendale","year":"2020","unstructured":"Mehendale, N.: Facial emotion recognition using convolutional neural networks (FERC). SN Appl. Sci. 2(3), 446 (2020). https:\/\/doi.org\/10.1007\/s42452-020-2234-1","journal-title":"SN Appl. Sci."},{"key":"4396_CR33","doi-asserted-by":"publisher","unstructured":"Wang, R., Guo, C., Shabaz, M., Rida, I., Cambria, E., Zhu, X.: CIME: contextual interaction-based multimodal emotion analysis with enhanced semantic information. IEEE Trans. Comput. Soc. Syst. 1\u201311 (2025). https:\/\/doi.org\/10.1109\/TCSS.2025.3572495","DOI":"10.1109\/TCSS.2025.3572495"},{"issue":"8","key":"4396_CR34","doi-asserted-by":"publisher","first-page":"374","DOI":"10.1007\/s40747-025-01931-8","volume":"11","author":"X Zhu","year":"2025","unstructured":"Zhu, X., Feng, H., Cambria, E., Huang, Y., Ju, M., Yuan, H., Wang, R.: EMVAS: end-to-end multimodal emotion visualization analysis system. Complex Intell. Syst. 11(8), 374 (2025). https:\/\/doi.org\/10.1007\/s40747-025-01931-8","journal-title":"Complex Intell. Syst."},{"issue":"5","key":"4396_CR35","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1109\/MIS.2025.3597120","volume":"40","author":"Y Zhang","year":"2025","unstructured":"Zhang, Y., Chen, H., Rida, I., Zhu, X.: A generative random modality dropout framework for robust multimodal emotion recognition. IEEE Intell. Syst. 40(5), 62\u201369 (2025). https:\/\/doi.org\/10.1109\/MIS.2025.3597120","journal-title":"IEEE Intell. Syst."},{"key":"4396_CR36","doi-asserted-by":"publisher","unstructured":"Wang, R., Wang, Y., Cambria, E., Fan, X., Yu, X., Huang, Y., E, X., Zhu, X.: Contrastive-based removal of negative information in multimodal emotion analysis. Cogn. Comput. 17(3), 107 (2025). https:\/\/doi.org\/10.1007\/s12559-025-10463-9","DOI":"10.1007\/s12559-025-10463-9"},{"issue":"4","key":"4396_CR37","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1007\/s13735-024-00347-3","volume":"13","author":"R Wang","year":"2024","unstructured":"Wang, R., Zhu, J., Wang, S., Wang, T., Huang, J., Zhu, X.: Multi-modal emotion recognition using tensor decomposition fusion and self-supervised multi-tasking. Int. J. Multimed. Inf. Retr. 13(4), 39 (2024). https:\/\/doi.org\/10.1007\/s13735-024-00347-3","journal-title":"Int. J. Multimed. Inf. Retr."},{"key":"4396_CR38","doi-asserted-by":"publisher","unstructured":"Tan, X., Xu, W., Wu, J., et al.: Au-guided feature aggregation for micro-expression recognition. Authorea (2025). https:\/\/doi.org\/10.22541\/au.174582245.51052169\/v1. Published April 28, 2025","DOI":"10.22541\/au.174582245.51052169\/v1"},{"key":"4396_CR39","doi-asserted-by":"publisher","first-page":"105721","DOI":"10.1016\/j.imavis.2025.105721","volume":"162","author":"J Chen","year":"2025","unstructured":"Chen, J., Xing, D., Shabaz, M., Zhu, Y., Wang, Y., Zhu, X.: DNLN: image super-resolution with deformable non-local attention and multi-branch weighted feature fusion. Image Vis. Comput. 162, 105721 (2025). https:\/\/doi.org\/10.1016\/j.imavis.2025.105721","journal-title":"Image Vis. Comput."},{"key":"4396_CR40","doi-asserted-by":"publisher","unstructured":"Wang, J., Gao, M., Zhai, W., Rida, I., Zhu, X., Li, Q.: Knowledge generation and distillation for road segmentation in intelligent transportation systems. IEEE Trans. Intell. Transp. Syst. 1\u201313 (2025). https:\/\/doi.org\/10.1109\/TITS.2025.3577794","DOI":"10.1109\/TITS.2025.3577794"},{"key":"4396_CR41","doi-asserted-by":"publisher","unstructured":"Wang, R., Zhou, K., Zhou, W., Shi, K., Pfaender, F., E, X., Zhu, X.: Multi-view residual spatio-temporal topology adaptive graph convolutional network for urban road traffic accident prediction with multi-source risks. Array 28, 100617 (2025). https:\/\/doi.org\/10.1016\/j.array.2025.100617","DOI":"10.1016\/j.array.2025.100617"},{"key":"4396_CR42","doi-asserted-by":"publisher","unstructured":"Sergeeva, A.D., Savin, A.V., Sablina, V.A., Melnik, O.V.: Emotion recognition from micro-expressions: search for the face and eyes. In: 2019 8th Mediterranean Conference on Embedded Computing (MECO), pp. 1\u20134 (2019). https:\/\/doi.org\/10.1109\/MECO.2019.8760029","DOI":"10.1109\/MECO.2019.8760029"},{"key":"4396_CR43","doi-asserted-by":"publisher","unstructured":"Priya, R.V., Bharat, R.: A novel geometric fuzzy membership functions for mouth and eye brows to recognize emotions. Concurr. Comput. Pract. Exp. 33(14) (2021). https:\/\/doi.org\/10.1002\/CPE.5610","DOI":"10.1002\/CPE.5610"},{"issue":"27","key":"4396_CR44","doi-asserted-by":"publisher","first-page":"42569","DOI":"10.1007\/S11042-023-14682-W","volume":"82","author":"SG Shaila","year":"2023","unstructured":"Shaila, S.G., Vadivel, A., Avani, V.S.: Emotion estimation from nose feature using pyramid structure. Multim. Tools Appl. 82(27), 42569\u201342591 (2023). https:\/\/doi.org\/10.1007\/S11042-023-14682-W","journal-title":"Multim. Tools Appl."},{"key":"4396_CR45","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1007\/978-3-031-43148-7_13","volume-title":"Image Analysis and Processing\u2014ICIAP 2023","author":"B De Carolis","year":"2023","unstructured":"De Carolis, B., Macchiarulo, N., Palestra, G., De Matteis, A.P., Lippolis, A.: FERmouth: facial emotion recognition from the mouth region. In: Foresti, G.L., Fusiello, A., Hancock, E. (eds.) Image Analysis and Processing\u2014ICIAP 2023, pp. 147\u2013158. Springer, Cham (2023)"},{"issue":"3","key":"4396_CR46","doi-asserted-by":"publisher","first-page":"1680","DOI":"10.3906\/ELK-1809-75","volume":"27","author":"J Bian","year":"2019","unstructured":"Bian, J., Mei, X., Xue, Y., Wu, L., Ding, Y.: Efficient hierarchical temporal segmentation method for facial expression sequences. Turkish J. Electr. Eng. Comput. Sci. 27(3), 1680\u20131695 (2019). https:\/\/doi.org\/10.3906\/ELK-1809-75","journal-title":"Turkish J. Electr. Eng. Comput. Sci."},{"key":"4396_CR47","doi-asserted-by":"publisher","unstructured":"Alazrai, R., Yousef, K.M.A., Daoud, M.I.: Emotion recognition based on decoupling the spatial context from the temporal dynamics of facial expressions. In: 2019 International Symposium on Networks, Computers and Communications (ISNCC), pp. 1\u20136 (2019). https:\/\/doi.org\/10.1109\/ISNCC.2019.8909141","DOI":"10.1109\/ISNCC.2019.8909141"},{"issue":"9","key":"4396_CR48","doi-asserted-by":"publisher","first-page":"14343","DOI":"10.1007\/S11042-020-10203-1","volume":"80","author":"J Wei","year":"2021","unstructured":"Wei, J., Yang, X., Dong, Y.: User-generated video emotion recognition based on key frames. Multim. Tools Appl. 80(9), 14343\u201314361 (2021). https:\/\/doi.org\/10.1007\/S11042-020-10203-1","journal-title":"Multim. Tools Appl."},{"key":"4396_CR49","doi-asserted-by":"publisher","unstructured":"Thuseethan, S., Rajasegarar, S., Yearwood, J.: Emotion intensity estimation from video frames using deep hybrid convolutional neural networks. In: 2019 International Joint Conference on Neural Networks (IJCNN), pp. 1\u201310 (2019). https:\/\/doi.org\/10.1109\/IJCNN.2019.8852365","DOI":"10.1109\/IJCNN.2019.8852365"},{"key":"4396_CR50","doi-asserted-by":"publisher","unstructured":"Lee, J., Lee, K., Park, H., Kim, I.-J., Nam, G.: V-NAW: video-based noise-aware adaptive weighting for facial expression recognition, pp. 5643\u20135650 (2025). https:\/\/doi.org\/10.1109\/CVPRW67362.2025.00561","DOI":"10.1109\/CVPRW67362.2025.00561"},{"issue":"1","key":"4396_CR51","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/TIP.2018.2868382","volume":"28","author":"S Li","year":"2019","unstructured":"Li, S., Deng, W.: Reliable crowdsourcing and deep locality-preserving learning for unconstrained facial expression recognition. IEEE Trans. Image Process. 28(1), 356\u2013370 (2019). https:\/\/doi.org\/10.1109\/TIP.2018.2868382","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"4396_CR52","first-page":"321","volume":"16","author":"NV Chawla","year":"2002","unstructured":"Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: Smote: synthetic minority over-sampling technique. J. Artif. Int. Res. 16(1), 321\u2013357 (2002)","journal-title":"J. Artif. Int. Res."},{"key":"4396_CR53","doi-asserted-by":"publisher","unstructured":"He, H., Bai, Y., Garcia, E.A., Li, S.: Adasyn: Adaptive synthetic sampling approach for imbalanced learning. In: 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), pp. 1322\u20131328 (2008). https:\/\/doi.org\/10.1109\/IJCNN.2008.4633969","DOI":"10.1109\/IJCNN.2008.4633969"},{"issue":"1","key":"4396_CR54","doi-asserted-by":"publisher","first-page":"861","DOI":"10.1007\/S10639-023-12342-Y","volume":"29","author":"Y Teng","year":"2024","unstructured":"Teng, Y., Wang, X.: Full caption, English proficiency and their relationships with behavioral, cognitive and emotional student engagement in Chinese EFL college content-based instruction video learning. Educ. Inf. Technol. 29(1), 861\u2013880 (2024). https:\/\/doi.org\/10.1007\/S10639-023-12342-Y","journal-title":"Educ. Inf. Technol."},{"issue":"2","key":"4396_CR55","doi-asserted-by":"publisher","first-page":"2128","DOI":"10.54364\/AAIML.2024.42122","volume":"4","author":"RY Ravenor","year":"2024","unstructured":"Ravenor, R.Y.: Ai-based facial emotion recognition solutions for education: a study of teacher-user and other categories. Adv. Artif. Intell. Mach. Learn. 4(2), 2128\u20132151 (2024). https:\/\/doi.org\/10.54364\/AAIML.2024.42122","journal-title":"Adv. Artif. Intell. Mach. Learn."},{"issue":"6","key":"4396_CR56","doi-asserted-by":"publisher","first-page":"521","DOI":"10.1080\/10447318.2018.1469710","volume":"35","author":"S Park","year":"2019","unstructured":"Park, S., Ryu, J.: Exploring preservice teachers\u2019 emotional experiences in an immersive virtual teaching simulation through facial expression recognition. Int. J. Hum. Comput. Interact. 35(6), 521\u2013533 (2019). https:\/\/doi.org\/10.1080\/10447318.2018.1469710","journal-title":"Int. J. Hum. Comput. Interact."},{"key":"4396_CR57","doi-asserted-by":"publisher","unstructured":"Tarchi, P., Gursesli, M.C., Cal\u00e0, F., Frassineti, L., Guazzini, A., Lanat\u00e0, A.: Markov chain modeling of facial emotions\u2019 dynamics in video games. In: 2024 IEEE International Workshop on Metrology for Industry 4.0 & IoT (MetroInd4.0 & IoT), pp. 33\u201337 (2024). https:\/\/doi.org\/10.1109\/MetroInd4.0IoT61288.2024.10584158","DOI":"10.1109\/MetroInd4.0IoT61288.2024.10584158"},{"key":"4396_CR58","doi-asserted-by":"publisher","unstructured":"Della\u00a0Greca, A., Ilaria, A., Tucci, C., Frugieri, N., Tortora, G.: A user study on the relationship between empathy and facial-based emotion simulation in virtual reality. In: Proceedings of the 2024 International Conference on Advanced Visual Interfaces. AVI \u201924. Association for Computing Machinery, New York, NY, USA (2024). https:\/\/doi.org\/10.1145\/3656650.3656691","DOI":"10.1145\/3656650.3656691"},{"issue":"3","key":"4396_CR59","doi-asserted-by":"publisher","first-page":"1717","DOI":"10.1007\/S10055-022-00720-9","volume":"27","author":"X Chen","year":"2023","unstructured":"Chen, X., Chen, H.: Emotion recognition using facial expressions in an immersive virtual reality application. Virtual Real. 27(3), 1717\u20131732 (2023). https:\/\/doi.org\/10.1007\/S10055-022-00720-9","journal-title":"Virtual Real."},{"issue":"12","key":"4396_CR60","doi-asserted-by":"publisher","first-page":"2607","DOI":"10.3390\/SYM14122607","volume":"14","author":"T Mazhar","year":"2022","unstructured":"Mazhar, T., Malik, M.A., Nadeem, M.A., Mohsan, S.A.H., Haq, I., Karim, F.K., Mostafa, S.M.: Movie reviews classification through facial image recognition and emotion detection using machine learning methods. Symmetry 14(12), 2607 (2022). https:\/\/doi.org\/10.3390\/SYM14122607","journal-title":"Symmetry"},{"key":"4396_CR61","doi-asserted-by":"publisher","unstructured":"Sahoo, G.K., Das, S.K., Singh, P.: Deep learning-based facial emotion recognition for driver healthcare. In: 2022 National Conference on Communications (NCC), pp. 154\u2013159 (2022). https:\/\/doi.org\/10.1109\/NCC55593.2022.9806751","DOI":"10.1109\/NCC55593.2022.9806751"},{"key":"4396_CR62","doi-asserted-by":"publisher","first-page":"77","DOI":"10.1016\/J.ARTMED.2019.06.004","volume":"98","author":"H Kalantarian","year":"2019","unstructured":"Kalantarian, H., Jedoui, K., Washington, P., Tariq, Q., Dunlap, K., Schwartz, J.N., Wall, D.P.: Labeling images with facial emotion and the potential for pediatric healthcare. Artif. Intell. Med. 98, 77\u201386 (2019). https:\/\/doi.org\/10.1016\/J.ARTMED.2019.06.004","journal-title":"Artif. Intell. Med."},{"issue":"1","key":"4396_CR63","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1007\/S10586-016-0535-3","volume":"19","author":"MA Alhussein","year":"2016","unstructured":"Alhussein, M.A.: Automatic facial emotion recognition using weber local descriptor for e-healthcare system. Clust. Comput. 19(1), 99\u2013108 (2016). https:\/\/doi.org\/10.1007\/S10586-016-0535-3","journal-title":"Clust. Comput."},{"key":"4396_CR64","doi-asserted-by":"publisher","unstructured":"Pham, P., Wang, J.: Understanding emotional responses to mobile video advertisements via physiological signal sensing and facial expression analysis. In: Proceedings of the 22nd International Conference on Intelligent User Interfaces. IUI \u201917, pp. 67\u201378. Association for Computing Machinery, New York, NY, USA (2017). https:\/\/doi.org\/10.1145\/3025171.3025186","DOI":"10.1145\/3025171.3025186"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-026-04396-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-026-04396-z","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-026-04396-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T13:39:07Z","timestamp":1775741947000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-026-04396-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3]]},"references-count":64,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["4396"],"URL":"https:\/\/doi.org\/10.1007\/s00371-026-04396-z","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3]]},"assertion":[{"value":"30 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"229"}}