{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T00:52:50Z","timestamp":1768956770765,"version":"3.49.0"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T00:00:00Z","timestamp":1768867200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T00:00:00Z","timestamp":1768867200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-026-21205-w","type":"journal-article","created":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T12:35:50Z","timestamp":1768912550000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Leveraging deep visual geometry group network for facial emotion recognition through RGB and thermal image fusion"],"prefix":"10.1007","volume":"85","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-0520-806X","authenticated-orcid":false,"given":"Tuan-Khoi","family":"Tran","sequence":"first","affiliation":[]},{"given":"Soo-Hyung","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Hyung-Jeong","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Seung-Won","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Ji-Eun","family":"Shin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,20]]},"reference":[{"key":"21205_CR1","doi-asserted-by":"publisher","first-page":"3859","DOI":"10.1109\/TMM.2021.3109419","volume":"24","author":"Y Pang","year":"2021","unstructured":"Pang Y, Lin J, Qin T, Chen Z (2021) Image-to-image translation: Methods and applications. IEEE Trans Multimed 24:3859\u20133881","journal-title":"IEEE Trans Multimed"},{"key":"21205_CR2","doi-asserted-by":"crossref","unstructured":"Lin J, Xia Y, Qin T, Chen Z, Liu T-Y (2018) Conditional image-to-image translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 5524\u20135532","DOI":"10.1109\/CVPR.2018.00579"},{"key":"21205_CR3","doi-asserted-by":"crossref","unstructured":"Shen Z, Huang M, Shi J, Xue X, Huang TS (2019) Towards instance-level image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 3683\u20133692","DOI":"10.1109\/CVPR.2019.00380"},{"key":"21205_CR4","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"key":"21205_CR5","doi-asserted-by":"crossref","unstructured":"Ko BC (2018) A brief review of facial emotion recognition based on visual information. sensors 18(2):401","DOI":"10.3390\/s18020401"},{"key":"21205_CR6","doi-asserted-by":"publisher","first-page":"593","DOI":"10.1016\/j.ins.2021.10.005","volume":"582","author":"FZ Canal","year":"2022","unstructured":"Canal FZ, M\u00fcller TR, Matias JC, Scotton GG, Sa Junior AR, Pozzebon E, Sobieranski AC (2022) A survey on facial emotion recognition techniques: A state-of-the-art literature review. Inf Sci 582:593\u2013617","journal-title":"Inf Sci"},{"key":"21205_CR7","doi-asserted-by":"crossref","unstructured":"Gursesli MC, Lombardi S, Duradoni M, Bocchi L, Guazzini A, Lanata A (2024) Facial emotion recognition (fer) through custom lightweight cnn model: performance evaluation in public datasets. IEEE Access","DOI":"10.1109\/ACCESS.2024.3380847"},{"issue":"1","key":"21205_CR8","first-page":"62","volume":"43","author":"A Ezquerra","year":"2025","unstructured":"Ezquerra A, Agen F, Toma RB, Ezquerra-Romano I (2025) Using facial emotion recognition to research emotional phases in an inquiry-based science activity. Res Sci Technol Educ 43(1):62\u201385","journal-title":"Res Sci Technol Educ"},{"key":"21205_CR9","doi-asserted-by":"crossref","unstructured":"L\u00f3pez-Morales H, Zabala M, Agulla L, Aguilar M, Sosa J, Vivas L, L\u00f3pez M (2025) Accuracy and speed in facial emotion recognition in children, adolescents, and adults. Curr Psychol 1\u201315","DOI":"10.1007\/s12144-025-07448-0"},{"issue":"1","key":"21205_CR10","first-page":"65","volume":"12","author":"M Pourmirzaei","year":"2025","unstructured":"Pourmirzaei M, Montazer GA, Mousavi E (2025) Attendee: an affective tutoring system based on facial emotion recognition and head pose estimation to personalize e-learning environment. J Comput Educ 12(1):65\u201392","journal-title":"J Comput Educ"},{"issue":"29","key":"21205_CR11","doi-asserted-by":"publisher","first-page":"73031","DOI":"10.1007\/s11042-024-18386-7","volume":"83","author":"VT-T Vo","year":"2024","unstructured":"Vo VT-T, Noh M-G, Kim S-H (2024) 4t-net: Multitask deep learning for nuclear analysis from pathology images. Multimed Tools Appl 83(29):73031\u201373053","journal-title":"Multimed Tools Appl"},{"key":"21205_CR12","unstructured":"Tran T-K, Kim S-H, Yang H-J, Kim S-W, Chen X (2024) Emotion recognition using rgb and thermal image fusion. In: In Proceeding of The 13th International Conference on Smart Media and Applications), p. 4. National University of Laos (Sokpaluang Campus), Vientiane, Laos"},{"key":"21205_CR13","doi-asserted-by":"crossref","unstructured":"Elsheikh RA, Mohamed M, Abou-Taleb AM, Ata MM (2025) Improving deep feature adequacy for facial emotion recognition: the impact of anti-aliasing on landmark-based and pixel-based approaches. Multimed Tools Appl 1\u201339","DOI":"10.1007\/s11042-025-20698-1"},{"key":"21205_CR14","doi-asserted-by":"crossref","unstructured":"Kwon B (2025) Data augmentation using convolutional autoencoder for facial emotion recognition. In: 2025 International Conference on Electronics, Information, and Communication (ICEIC), pp 1\u20134. IEEE","DOI":"10.1109\/ICEIC64972.2025.10879763"},{"key":"21205_CR15","doi-asserted-by":"crossref","unstructured":"Puthanidam, R.V., Moh, T.-S.: A hybrid approach for facial expression recognition. In: Proceedings of the 12th International Conference on Ubiquitous Information Management and Communication, pp. 1\u20138 (2018)","DOI":"10.1145\/3164541.3164593"},{"key":"21205_CR16","doi-asserted-by":"publisher","first-page":"71311","DOI":"10.1109\/ACCESS.2022.3188730","volume":"10","author":"T Dar","year":"2022","unstructured":"Dar T, Javed A, Bourouis S, Hussein HS, Alshazly H (2022) Efficient-swishnet based system for facial emotion recognition. IEEE Access 10:71311\u201371328","journal-title":"IEEE Access"},{"key":"21205_CR17","doi-asserted-by":"publisher","first-page":"306","DOI":"10.1016\/j.neucom.2021.10.038","volume":"468","author":"Z Fei","year":"2022","unstructured":"Fei Z, Yang E, Yu L, Li X, Zhou H, Zhou W (2022) A novel deep neural network-based emotion analysis system for automatic detection of mild cognitive impairment in the elderly. Neurocomputing 468:306\u2013316","journal-title":"Neurocomputing"},{"key":"21205_CR18","doi-asserted-by":"publisher","first-page":"52509","DOI":"10.1109\/ACCESS.2021.3069881","volume":"9","author":"VG Mahesh","year":"2021","unstructured":"Mahesh VG, Chen C, Rajangam V, Raj ANJ, Krishnan PT (2021) Shape and texture aware facial expression recognition using spatial pyramid zernike moments and law\u2019s textures feature set. IEEE Access 9:52509\u201352522","journal-title":"IEEE Access"},{"issue":"10","key":"21205_CR19","doi-asserted-by":"publisher","first-page":"951","DOI":"10.1111\/psyp.12243","volume":"51","author":"S Ioannou","year":"2014","unstructured":"Ioannou S, Gallese V, Merla A (2014) Thermal infrared imaging in psychophysiology: potentialities and limits. Psychophysiology 51(10):951\u2013963. https:\/\/doi.org\/10.1111\/psyp.12243","journal-title":"Psychophysiology"},{"issue":"1","key":"21205_CR20","doi-asserted-by":"publisher","first-page":"530","DOI":"10.1109\/TAFFC.2019.2946774","volume":"13","author":"P Werner","year":"2022","unstructured":"Werner P, Lopez-Martinez D, Walter S, Al-Hamadi A, Gruss S, Picard RW (2022) Automatic recognition methods supporting pain assessment: A survey. IEEE Trans Affect Comput 13(1):530\u2013552. https:\/\/doi.org\/10.1109\/TAFFC.2019.2946774","journal-title":"IEEE Trans Affect Comput"},{"key":"21205_CR21","doi-asserted-by":"crossref","unstructured":"Ruiz-del-Solar J, Verschae R, Hermosilla G, Correa M (2013) Thermal face recognition in unconstrained environments using histograms of lbp features. In: Local Binary Patterns: New Variants and Applications, pp 219\u2013243. Springer, ???","DOI":"10.1007\/978-3-642-39289-4_10"},{"key":"21205_CR22","doi-asserted-by":"crossref","unstructured":"Ilikci B, Chen L, Cho H, Liu Q (2019) Heat-map based emotion and face recognition from thermal images. In: 2019 Computing, Communications and IoT Applications (ComComAp), pp 449\u2013453. IEEE","DOI":"10.1109\/ComComAp46287.2019.9018786"},{"key":"21205_CR23","doi-asserted-by":"crossref","unstructured":"Tian P et al (2023) Multimodal emotion detection based on visual and thermal images. In: ACM International Conference on Multimodal Interaction (ICMI)","DOI":"10.1145\/3599589.3599590"},{"key":"21205_CR24","doi-asserted-by":"crossref","unstructured":"Adra M, Mirabet-Herranz N, Dugelay J-L (2024) Tri-modal microexpression recognition with rgb, thermal, and event data. In: Proceedings of the 17th International Conference on PErvasive Technologies Related to Assistive Environments","DOI":"10.1007\/978-3-031-87660-8_23"},{"key":"21205_CR25","doi-asserted-by":"crossref","unstructured":"Yu C, Tapus A (2020) Multimodal emotion recognition with thermal and rgb-d cameras for human-robot interaction. In: Proceedings of the 25th International Conference on Intelligent User Interfaces","DOI":"10.1145\/3371382.3378342"},{"key":"21205_CR26","doi-asserted-by":"crossref","unstructured":"Fan C-M, Liu T-J, Liu K-H (2022) Sunet: Swin transformer unet for image denoising. In: 2022 IEEE International Symposium on Circuits and Systems (ISCAS), pp 2333\u20132337. IEEE","DOI":"10.1109\/ISCAS48785.2022.9937486"},{"key":"21205_CR27","doi-asserted-by":"crossref","unstructured":"Kuzdeuov A, Koishigarina D, Aubakirova D, Abushakimova S, Varol HA (2022) Sf-tl54: A thermal facial landmark dataset with visual pairs. In: 2022 IEEE\/SICE International Symposium on System Integration (SII), pp 748\u2013753. IEEE","DOI":"10.1109\/SII52469.2022.9708901"},{"key":"21205_CR28","volume-title":"The Karolinska Directed Emotional Faces","author":"D Lundqvist","year":"1998","unstructured":"Lundqvist D, Flykt A, \u00d6hman A (1998) The Karolinska Directed Emotional Faces. KDEF), Stockholm, Sweden"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21205-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-026-21205-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21205-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T12:35:54Z","timestamp":1768912554000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-026-21205-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,20]]},"references-count":28,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2026,1]]}},"alternative-id":["21205"],"URL":"https:\/\/doi.org\/10.1007\/s11042-026-21205-w","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,20]]},"assertion":[{"value":"5 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 September 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 November 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 January 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"We confirm that all participants have been fully informed about the purpose of the study and the procedures involved.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"We confirm that consent has been obtained from all individuals included in this publication. All necessary permissions have been secured to use the relevant data, images, and information. The individuals involved have been fully informed about the publication and have agreed to the dissemination of their information.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to publish"}},{"value":"The authors declare that they have no conflict of interest.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"11"}}