{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,29]],"date-time":"2025-04-29T11:10:02Z","timestamp":1745925002390,"version":"3.40.4"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,4,11]],"date-time":"2025-04-11T00:00:00Z","timestamp":1744329600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,4,11]],"date-time":"2025-04-11T00:00:00Z","timestamp":1744329600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62362030"],"award-info":[{"award-number":["62362030"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the Jiangxi Province Graduate Innovation Special Fund Project","award":["No. YC2023-s790"],"award-info":[{"award-number":["No. YC2023-s790"]}]},{"DOI":"10.13039\/501100004479","name":"Natural Science Foundation of Jiangxi Province","doi-asserted-by":"publisher","award":["No.20224ACB202011"],"award-info":[{"award-number":["No.20224ACB202011"]}],"id":[{"id":"10.13039\/501100004479","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,5]]},"DOI":"10.1007\/s40747-025-01855-3","type":"journal-article","created":{"date-parts":[[2025,4,11]],"date-time":"2025-04-11T09:22:02Z","timestamp":1744363322000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Micro-expression spotting based on multi-modal hierarchical semantic guided deep fusion and optical flow driven feature integration"],"prefix":"10.1007","volume":"11","author":[{"given":"Haolin","family":"Chang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4226-7497","authenticated-orcid":false,"given":"Zhihua","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Fan","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,11]]},"reference":[{"key":"1855_CR1","doi-asserted-by":"publisher","unstructured":"Davison A, Merghani W, Lansley C, et al (2018) Objective micro-facial movement detection using facs-based regions and baseline evaluation. In: 2018 13th IEEE international conference on automatic face & gesture recognition (FG 2018). IEEE, pp. 642\u2013649. https:\/\/doi.org\/10.1109\/FG.2018.00101","DOI":"10.1109\/FG.2018.00101"},{"key":"1855_CR2","doi-asserted-by":"publisher","first-page":"338","DOI":"10.1037\/a0026118","volume":"2","author":"ME Kemeny","year":"2012","unstructured":"Kemeny ME, Foltz C, Cavanagh JF et al (2012) Contemplative emotion training reduces negative emotional behavior and promotes prosocial responses. Emotion 2:338\u2013350. https:\/\/doi.org\/10.1037\/a0026118","journal-title":"Emotion"},{"key":"1855_CR3","unstructured":"Ekman P (2009) Telling lies: clues to deceit in the marketplace, politics, and marriage (revised edition). WW Norton and Company"},{"key":"1855_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2024.3522423","volume":"74","author":"B Qi","year":"2025","unstructured":"Qi B, Li Q, Zhang Y et al (2025) Infrared and visible image fusion via sparse representation and adaptive dual-channel PCNN model based on co-occurrence analysis shearlet transform. IEEE Trans Instrum Meas 74:1\u201315. https:\/\/doi.org\/10.1109\/TIM.2024.3522423","journal-title":"IEEE Trans Instrum Meas"},{"key":"1855_CR5","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2024.3524794","author":"Z Huang","year":"2025","unstructured":"Huang Z, Lin C, Xu B et al (2025) T2EA: target-aware taylor expansion approximation network for infrared and visible image fusion. IEEE Trans Circ Syst Video Technol. https:\/\/doi.org\/10.1109\/TCSVT.2024.3524794","journal-title":"IEEE Trans Circ Syst Video Technol"},{"key":"1855_CR6","doi-asserted-by":"publisher","unstructured":"Stewart P A, Waller B M, Schubert J N (2009) Presidential speechmaking style: Emotional response to micro-expressions of facial affect. Motiv Emot 33:125\u2013135. https:\/\/doi.org\/10.1007\/s11031-009-9129-1","DOI":"10.1007\/s11031-009-9129-1"},{"key":"1855_CR7","doi-asserted-by":"publisher","first-page":"530","DOI":"10.1007\/s10979-008-9166-4","volume":"33","author":"M O\u2019sullivan","year":"2009","unstructured":"O\u2019sullivan M, Frank MG, Hurley CM et al (2009) Police lie detection accuracy: the effect of lie scenario. Law Hum Behav 33:530\u2013538. https:\/\/doi.org\/10.1007\/s10979-008-9166-4","journal-title":"Law Hum Behav"},{"key":"1855_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/1472-6920-9-47","volume":"9","author":"J Endres","year":"2009","unstructured":"Endres J, Laidlaw A (2009) Micro-expression recognition training in medical students: a pilot study. BMC Med Educ 9:1\u20136. https:\/\/doi.org\/10.1186\/1472-6920-9-47","journal-title":"BMC Med Educ"},{"key":"1855_CR9","doi-asserted-by":"publisher","unstructured":"Wang L, Huang P, Cai W, et al (2024) Micro-expression recognition by fusing action unit detection and Spatio-temporal features. In: ICASSP 2024\u20132024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp. 5595\u20135599. https:\/\/doi.org\/10.1109\/ICASSP48485.2024.10446702","DOI":"10.1109\/ICASSP48485.2024.10446702"},{"key":"1855_CR10","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2024.3520499","author":"J Qi","year":"2025","unstructured":"Qi J, Ni B, Yu Q et al (2025) EPAFusion: a novel fusion network based on enhancement and progressive aware for infrared\u2013visible images in low-light. IEEE Sens J. https:\/\/doi.org\/10.1109\/JSEN.2024.3520499","journal-title":"IEEE Sens J"},{"key":"1855_CR11","doi-asserted-by":"publisher","unstructured":"Zhang H, Xu H, Xiao Y, et al (2020) Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In: the AAAI conference on artificial intelligence, pp 12797\u201312804. https:\/\/doi.org\/10.1609\/aaai.v34i07.6975","DOI":"10.1609\/aaai.v34i07.6975"},{"key":"1855_CR12","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1016\/j.inffus.2023.02.027","volume":"95","author":"J Liu","year":"2023","unstructured":"Liu J, Wu G, Luan J et al (2023) HoLoCo: holistic and local contrastive learning network for multi-exposure image fusion. Inform Fus 95:237\u2013249. https:\/\/doi.org\/10.1016\/j.inffus.2023.02.027","journal-title":"Inform Fus"},{"issue":"4","key":"1855_CR13","doi-asserted-by":"publisher","first-page":"563","DOI":"10.1109\/TAFFC.2017.2667642","volume":"9","author":"X Li","year":"2017","unstructured":"Li X, Hong X, Moilanen A, Huang X et al (2017) Towards reading hidden emotions: a comparative study of spontaneous micro-expression spotting and recognition methods. IEEE Trans Affect Comput 9(4):563\u2013577. https:\/\/doi.org\/10.1109\/TAFFC.2017.2667642","journal-title":"IEEE Trans Affect Comput"},{"issue":"2","key":"1855_CR14","doi-asserted-by":"publisher","first-page":"458","DOI":"10.31577\/cai_2024_2_458","volume":"43","author":"B Zhang","year":"2024","unstructured":"Zhang B, Lu J, Wang C et al (2024) FESNet: spotting facial expressions using local spatial discrepancy and multi-scale temporal aggregation. Comput Inform 43(2):458\u2013481","journal-title":"Comput Inform"},{"key":"1855_CR15","unstructured":"Lu H, Kpalma K, Ronsin J (2017) Micro-expression detection using integral projections. J WSCG 25(2):87\u201396. http:\/\/hdl.handle.net\/11025\/26286"},{"key":"1855_CR16","doi-asserted-by":"publisher","DOI":"10.1109\/TCE.2024.3474037","author":"Y Song","year":"2024","unstructured":"Song Y, Liu Y, Lin Z et al (2024) Learning from AI-generated annotations for medical image segmentation. IEEE Trans Consum Electron. https:\/\/doi.org\/10.1109\/TCE.2024.3474037","journal-title":"IEEE Trans Consum Electron"},{"key":"1855_CR17","doi-asserted-by":"publisher","first-page":"170","DOI":"10.1016\/j.image.2016.06.004","volume":"47","author":"ST Liong","year":"2016","unstructured":"Liong ST, See J, Phan R et al (2016) Spontaneous subtle expression detection and recognition based on facial strain. Signal Process Image Commun 47:170\u2013182. https:\/\/doi.org\/10.1016\/j.image.2016.06.004","journal-title":"Signal Process Image Commun"},{"key":"1855_CR18","doi-asserted-by":"publisher","DOI":"10.1109\/TCE.2024.3476129","author":"D Cai","year":"2024","unstructured":"Cai D, Chen K, Lin Z et al (2024) JointSTNet: joint pre-training for spatial-temporal traffic forecasting. IEEE Trans Consum Electron. https:\/\/doi.org\/10.1109\/TCE.2024.3476129","journal-title":"IEEE Trans Consum Electron"},{"key":"1855_CR19","doi-asserted-by":"publisher","first-page":"3956","DOI":"10.1109\/TIP.2021.3064258","volume":"30","author":"SJ Wang","year":"2021","unstructured":"Wang SJ, He Y, Li J et al (2021) MESNet: a convolutional neural network for spotting multi-scale micro-expression intervals in long videos. IEEE Trans Image Process 30:3956\u20133969. https:\/\/doi.org\/10.1109\/TIP.2021.3064258","journal-title":"IEEE Trans Image Process"},{"key":"1855_CR20","doi-asserted-by":"publisher","unstructured":"Li X, Li J, Du W, et al (2024) Learning interval-aware embedding for macro-and micro-expression spotting. In: the Asian Conference on Computer Vision, pp. 337\u2013353. https:\/\/doi.org\/10.1109\/ICSP.2016.7878004","DOI":"10.1109\/ICSP.2016.7878004"},{"key":"1855_CR21","doi-asserted-by":"publisher","first-page":"76833","DOI":"10.1007\/s11042-024-18645-7","volume":"83","author":"Z Xie","year":"2024","unstructured":"Xie Z, Fan J, Cheng S (2024) Multi-channel capsule network for micro-expression recognition with multiscale fusion. Multimed Tools Appl 83:76833\u201376850. https:\/\/doi.org\/10.1007\/s11042-024-18645-7","journal-title":"Multimed Tools Appl"},{"key":"1855_CR22","doi-asserted-by":"publisher","unstructured":"Nag S, Bhunia A K, Konwer A, et al (2019) Facial micro-expression spotting and recognition using time contrasted feature with visual memory. In: ICASSP 2019\u20132019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp 2022\u20132026. https:\/\/doi.org\/10.1109\/ICASSP.2019.8683737","DOI":"10.1109\/ICASSP.2019.8683737"},{"key":"1855_CR23","doi-asserted-by":"publisher","unstructured":"Zhang Z, Zhao S, Mao X, et al (2024) A multi-scale feature learning network with optical flow correction for micro-and macro-expression spotting. In: the 32nd ACM International Conference on Multimedia, pp. 11497\u201311502. https:\/\/doi.org\/10.1145\/3664647.3689143","DOI":"10.1145\/3664647.3689143"},{"key":"1855_CR24","doi-asserted-by":"publisher","unstructured":"Yu J, Zhao G, Zhang Y et al (2024) Temporal-informative adapters in VideoMAE V2 and multi-scale feature fusion for micro-expression spotting-then-recognize. In: the 32nd ACM International Conference on Multimedia, pp. 11484\u201311489. https:\/\/doi.org\/10.1145\/3664647.3689141","DOI":"10.1145\/3664647.3689141"},{"key":"1855_CR25","doi-asserted-by":"publisher","unstructured":"Yang B, Wu J, Zhou Z, et al (2021) Facial action unit-based deep learning framework for spotting macro-and micro-expressions in long video sequences. In: the 29th ACM International Conference on Multimedia. ACM, pp. 4794\u20134798. https:\/\/doi.org\/10.1145\/3474085.3479209","DOI":"10.1145\/3474085.3479209"},{"key":"1855_CR26","doi-asserted-by":"publisher","unstructured":"Yap C H, Kendrick C, Yap M H (2020) Samm long videos: a spontaneous facial micro-and macro-expressions dataset. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020). IEEE, pp. 771\u2013776. https:\/\/doi.org\/10.1109\/FG47880.2020.00029","DOI":"10.1109\/FG47880.2020.00029"},{"key":"1855_CR27","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen J, Li X, Luo L et al (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64\u201378. https:\/\/doi.org\/10.1016\/j.ins.2019.08.066","journal-title":"Inf Sci"},{"key":"1855_CR28","doi-asserted-by":"publisher","unstructured":"Yu W W, Jiang J, Li Y J (2021) LSSNet: a two-stream convolutional neural network for spotting macro-and micro-expression in long videos. In: the 29th ACM International Conference on Multimedia, ACM, pp. 4745\u20134749. https:\/\/doi.org\/10.1145\/3474085.3479215","DOI":"10.1145\/3474085.3479215"},{"key":"1855_CR29","doi-asserted-by":"publisher","first-page":"178","DOI":"10.1016\/j.neunet.2024.106421","volume":"2024","author":"H Tang","year":"2024","unstructured":"Tang H, Chai L (2024) Facial micro-expression recognition using stochastic graph convolutional network and dual transferred learning. Neural Netw 2024:178. https:\/\/doi.org\/10.1016\/j.neunet.2024.106421","journal-title":"Neural Netw"},{"key":"1855_CR30","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li H, Wu XJ, Kittler J (2020) MDLatLRR: a novel decomposition method for infrared and visible image fusion. IEEE Trans Image Process 29:4733\u20134746. https:\/\/doi.org\/10.1109\/TIP.2020.2975984","journal-title":"IEEE Trans Image Process"},{"key":"1855_CR31","doi-asserted-by":"publisher","first-page":"461","DOI":"10.3390\/electronics13020461","volume":"13","author":"Z Xie","year":"2024","unstructured":"Xie Z, Zhao C (2024) Dual-branch cross-attention network for micro-expression recognition with transformer variants. Electronics 13:461. https:\/\/doi.org\/10.3390\/electronics13020461","journal-title":"Electronics"},{"issue":"5","key":"1855_CR32","doi-asserted-by":"publisher","first-page":"743","DOI":"10.1109\/JSEN.2007.894926","volume":"7","author":"N Cvejic","year":"2007","unstructured":"Cvejic N, Bull D, Canagarajah N (2007) Region-based multimodal image fusion using ICA bases. IEEE Sens J 7(5):743\u2013751. https:\/\/doi.org\/10.1109\/JSEN.2007.894926","journal-title":"IEEE Sens J"},{"key":"1855_CR33","doi-asserted-by":"publisher","first-page":"70","DOI":"10.1016\/j.neucom.2012.12.015","volume":"111","author":"J Han","year":"2013","unstructured":"Han J, Pauwels EJ, De Zeeuw P (2013) Fast saliency-aware multi-modality image fusion. Neurocomputing 111:70\u201380. https:\/\/doi.org\/10.1016\/j.neucom.2012.12.015","journal-title":"Neurocomputing"},{"key":"1855_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2024.105410","volume":"2025","author":"Y Fu","year":"2025","unstructured":"Fu Y, Liu Z, Peng J et al (2025) GANSD: a generative adversarial network based on saliency detection for infrared and visible image fusion. Image Vis Comput 2025:105410. https:\/\/doi.org\/10.1016\/j.imavis.2024.105410","journal-title":"Image Vis Comput"},{"key":"1855_CR35","doi-asserted-by":"publisher","unstructured":"Li Y, Huang X, Zhao G (2018) Can micro-expression be recognized based on single apex frame? In: 2018 25th IEEE International Conference on Image Processing (ICIP). IEEE, pp. 3094\u20133098. https:\/\/doi.org\/10.1109\/ICIP.2018.8451376","DOI":"10.1109\/ICIP.2018.8451376"},{"key":"1855_CR36","doi-asserted-by":"publisher","first-page":"130","DOI":"10.1016\/j.patrec.2024.01.004","volume":"178","author":"J Zhou","year":"2024","unstructured":"Zhou J, Wu Y (2024) Micro-expression spotting with a novel wavelet convolution magnification network in long videos. Pattern Recogn Lett 178:130\u2013137. https:\/\/doi.org\/10.1016\/j.patrec.2024.01.004","journal-title":"Pattern Recogn Lett"},{"key":"1855_CR37","doi-asserted-by":"publisher","unstructured":"Deng X, Yang R, Xu M, et al (2019) Wavelet domain style transfer for an effective perception-distortion tradeoff in single image super-resolution. In: the IEEE\/CVF international conference on computer vision, pp. 3076\u20133085. https:\/\/doi.org\/10.48550\/arXiv.1910.04074","DOI":"10.48550\/arXiv.1910.04074"},{"key":"1855_CR38","doi-asserted-by":"publisher","unstructured":"Shreve M, Godavarthy S, Manohar V et al (2009) Towards macro-and micro-expression spotting in video using strain patterns. In: 2009 Workshop on Applications of Computer Vision (WACV). IEEE, pp. 1\u20136. https:\/\/doi.org\/10.1109\/WACV.2009.5403044","DOI":"10.1109\/WACV.2009.5403044"},{"key":"1855_CR39","doi-asserted-by":"publisher","unstructured":"Yu J, Zhang Y, Zhao G et al (2024) Micro-expression spotting based on optical flow feature with boundary calibration. In: the 32nd ACM International Conference on Multimedia. IEEE, pp. 11490\u201311496. https:\/\/doi.org\/10.1145\/3664647.3689142","DOI":"10.1145\/3664647.3689142"},{"issue":"18","key":"1855_CR40","doi-asserted-by":"publisher","first-page":"3947","DOI":"10.3390\/electronics12183947","volume":"12","author":"W Fu","year":"2023","unstructured":"Fu W, An Z, Huang W et al (2023) A spatio-temporal spotting network with sliding windows for micro-expression detection. Electronics 12(18):3947\u20133951. https:\/\/doi.org\/10.3390\/electronics12183947","journal-title":"Electronics"},{"key":"1855_CR41","doi-asserted-by":"publisher","unstructured":"Bai M, Goecke R (2024) Can expression sensitivity improve macro-and micro-expression spotting in long videos? In: the 2nd International Workshop on Multimodal and Responsible Affective Computing. IEEE, pp. 30\u201338. https:\/\/doi.org\/10.1145\/3689092.368939","DOI":"10.1145\/3689092.368939"},{"key":"1855_CR42","doi-asserted-by":"publisher","unstructured":"Xie Z, Ye X (2024) Local and global features interactive fusion network for macro-and micro-expression spotting in long videos. In: Chinese Conference on Pattern Recognition and Computer Vision (PRCV). Springer, pp 336\u2013350. https:\/\/doi.org\/10.1007\/978-981-97-8795-1_23","DOI":"10.1007\/978-981-97-8795-1_23"},{"issue":"2","key":"1855_CR43","doi-asserted-by":"publisher","first-page":"458","DOI":"10.31577\/cai_2024_2_458","volume":"43","author":"B Zhang","year":"2024","unstructured":"Zhang B, Lu J, Wang C et al (2024) FESNet: spotting facial expressions using local spatial discrepancy and multi-scale temporal aggregation. Comput Inform 43(2):458\u2013481. https:\/\/doi.org\/10.31577\/cai_2024_2_458","journal-title":"Comput Inform"},{"key":"1855_CR44","doi-asserted-by":"publisher","unstructured":"Yap C H, Yap M H, Davison A, et al (2022) 3d-cnn for facial micro-and macro-expression spotting on long video sequences using temporal oriented reference frame. In: the 30th ACM International Conference on Multimedia. ACM, pp. 7016\u20137020. https:\/\/doi.org\/10.1145\/3503161.3551570","DOI":"10.1145\/3503161.3551570"},{"key":"1855_CR45","doi-asserted-by":"publisher","unstructured":"He Y, Liu W, Wang G et al (2024) Enhancing micro-expression analysis performance by effectively addressing data imbalance. In: the 32nd ACM International Conference on Multimedia. ACM, pp. 11503\u201311507. https:\/\/doi.org\/10.1145\/3664647.3689144","DOI":"10.1145\/3664647.3689144"},{"key":"1855_CR46","doi-asserted-by":"publisher","first-page":"146","DOI":"10.1016\/j.patrec.2023.03.012","volume":"168","author":"X Guo","year":"2023","unstructured":"Guo X, Zhang X, Li L et al (2023) Micro-expression spotting with multi-scale local transformer in long videos. Pattern Recogn Lett 168:146\u2013152. https:\/\/doi.org\/10.1016\/j.patrec.2023.03.012","journal-title":"Pattern Recogn Lett"},{"issue":"10","key":"1855_CR47","doi-asserted-by":"publisher","first-page":"2143","DOI":"10.3724\/SP.J.1042.2022.02143","volume":"30","author":"J Li","year":"2022","unstructured":"Li J, Dong Z, Liu Y et al (2022) Micro-expression spotting method based on human attention mechanism. Adv Psychol Sci 30(10):2143\u20132153. https:\/\/doi.org\/10.3724\/SP.J.1042.2022.02143","journal-title":"Adv Psychol Sci"},{"key":"1855_CR48","doi-asserted-by":"publisher","unstructured":"Deng Y, Hayashi H, Nagahara H (2024) Multi-scale spatio-temporal graph convolutional network for facial expression spotting. In: 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition (FG). IEEE, pp. 1\u201310. https:\/\/doi.org\/10.48550\/arXiv.2403.15994","DOI":"10.48550\/arXiv.2403.15994"},{"key":"1855_CR49","doi-asserted-by":"publisher","unstructured":"Yu WW, Zhang XS, Luo FY, et al. (2024) Weak supervision with arbitrary single frame for micro-and macro-expression spotting. https:\/\/doi.org\/10.48550\/arXiv.2403.14240. arXiv preprint arXiv:2403.14240","DOI":"10.48550\/arXiv.2403.14240"},{"key":"1855_CR50","doi-asserted-by":"publisher","unstructured":"Wang L, Huang P, Cai W, et al (2024) Micro-expression recognition by fusing action unit detection and Spatio-temporal features. In: ICASSP 2024\u20132024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp. 5595\u20135599. https:\/\/doi.org\/10.1109\/ICASSP48485.2024.10446702","DOI":"10.1109\/ICASSP48485.2024.10446702"},{"key":"1855_CR51","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2023.3273451","author":"D Rao","year":"2023","unstructured":"Rao D, Xu T, Wu X (2023) TGFuse: an infrared and visible image fusion approach based on transformer and generative adversarial network. IEEE Trans Image Process. https:\/\/doi.org\/10.1109\/TIP.2023.3273451","journal-title":"IEEE Trans Image Process"},{"key":"1855_CR52","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3191664","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang Z, Chen Y, Shao W et al (2022) SwinFuse: a residual swin transformer fusion network for infrared and visible images. IEEE Trans Instrum Measur 71:1\u201312. https:\/\/doi.org\/10.1109\/TIM.2022.3191664","journal-title":"IEEE Trans Instrum Measur"},{"key":"1855_CR53","doi-asserted-by":"publisher","unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. Advances in neural information processing systems. ACM, pp. 6000\u20136010. https:\/\/doi.org\/10.48550\/arXiv.1706.03762","DOI":"10.48550\/arXiv.1706.03762"},{"key":"1855_CR54","doi-asserted-by":"publisher","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A et al (2020) An image is worth 16x16 words: Transformers for image recognition at scale. https:\/\/doi.org\/10.48550\/arXiv.2010.11929. arXiv preprint arXiv:2010.11929","DOI":"10.48550\/arXiv.2010.11929"},{"key":"1855_CR55","doi-asserted-by":"publisher","unstructured":"Liu Z, Lin Y, Cao Y, et al (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: the IEEE\/CVF international conference on computer vision. IEEE, pp. 10012\u201310022. https:\/\/doi.org\/10.48550\/arXiv.2103.14030","DOI":"10.48550\/arXiv.2103.14030"},{"issue":"1","key":"1855_CR56","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1007\/s11760-024-03800-2","volume":"19","author":"M Tao","year":"2025","unstructured":"Tao M, Long H, Kong G et al (2025) Combining information augmentation aggregation and dual-granularity feature fusion for visible-infrared person re-identification. SIViP 19(1):163. https:\/\/doi.org\/10.1007\/s11760-024-03800-2","journal-title":"SIViP"},{"key":"1855_CR57","doi-asserted-by":"publisher","unstructured":"Nguyen X B, Duong C N, Li X et al. (2023) Micron-bert: bert-based facial micro-expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, pp.1482\u20131492. https:\/\/doi.org\/10.48550\/arXiv.2304.03195","DOI":"10.48550\/arXiv.2304.03195"},{"key":"1855_CR58","unstructured":"Lok K (2024) A proficient approach utilising high-level representations for micro-expression spotting. In: Tunku Abdul Rahman University of Management and Technology. https:\/\/eprints.tarc.edu.my\/id\/eprint\/31298."},{"issue":"4","key":"1855_CR59","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1109\/TAFFC.2017.2654440","volume":"9","author":"F Qu","year":"2017","unstructured":"Qu F, Wang SJ, Yan WJ et al (2017) CAS (ME)2: a database for spontaneous macro-expression and micro-expression spotting and recognition. IEEE Trans Affect Comput 9(4):424\u2013436. https:\/\/doi.org\/10.1109\/TAFFC.2017.2654440","journal-title":"IEEE Trans Affect Comput"},{"key":"1855_CR60","doi-asserted-by":"publisher","unstructured":"Yap C, Kendrick C, Yap M (2020) Samm long videos: a spontaneous facial micro-and macro-expressions dataset. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020). IEEE, pp. 771\u2013776. https:\/\/doi.org\/10.1109\/FG47880.2020.00029","DOI":"10.1109\/FG47880.2020.00029"},{"key":"1855_CR61","doi-asserted-by":"publisher","unstructured":"Liong G B, Liong S T, See J, et al. (2022) Mtsn: a multi-temporal stream network for spotting facial macro-and micro-expression with hard and soft pseudo-labels. In: the 2nd Workshop on Facial Micro-Expression: Advanced Techniques for Multi-Modal Facial Expression Analysis, pp. 3\u201310. https:\/\/doi.org\/10.1145\/3552465.3555040","DOI":"10.1145\/3552465.3555040"},{"key":"1855_CR62","doi-asserted-by":"publisher","unstructured":"He Y, Wang S J, Li J et al. (2020) Spotting macro-and micro-expression intervals in long video sequences. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020). IEEE, pp. 742\u2013748. https:\/\/doi.org\/10.1109\/FG47880.2020.00036","DOI":"10.1109\/FG47880.2020.00036"},{"key":"1855_CR63","doi-asserted-by":"publisher","unstructured":"Zhang L, Li J, Wang S et al. (2020) Spatio-temporal fusion for macro-and micro-expression spotting in long video sequences. In: 15th IEEE international conference on automatic face and gesture recognition (FG 2020). IEEE, pp:734\u2013741. https:\/\/doi.org\/10.1109\/FG47880.2020.00037","DOI":"10.1109\/FG47880.2020.00037"},{"key":"1855_CR64","doi-asserted-by":"publisher","unstructured":"Yin S, Wu S, Xu T et al. (2023) AU-aware graph convolutional network for Macro and Micro-expression spotting. In: 2023 IEEE International Conference on Multimedia and Expo (ICME). IEEE, pp 228\u2013233. https:\/\/doi.org\/10.1109\/ICME55011.2023.00047","DOI":"10.1109\/ICME55011.2023.00047"},{"key":"1855_CR65","doi-asserted-by":"publisher","unstructured":"Selvaraju RR, Cogswell M, Das A, et al. (2017) Grad-cam: visual explanations from deep networks via gradient-based localization. In: the IEEE international conference on computer vision. IEEE, pp. 618\u2013626. https:\/\/doi.org\/10.1007\/s11263-019-01228-7","DOI":"10.1007\/s11263-019-01228-7"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01855-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-01855-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01855-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,29]],"date-time":"2025-04-29T10:37:45Z","timestamp":1745923065000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-01855-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,11]]},"references-count":65,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,5]]}},"alternative-id":["1855"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-01855-3","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"type":"print","value":"2199-4536"},{"type":"electronic","value":"2198-6053"}],"subject":[],"published":{"date-parts":[[2025,4,11]]},"assertion":[{"value":"12 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 March 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 April 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"238"}}