{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T11:48:56Z","timestamp":1773229736547,"version":"3.50.1"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T00:00:00Z","timestamp":1765584000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T00:00:00Z","timestamp":1769040000000},"content-version":"vor","delay-in-days":40,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s44443-025-00414-7","type":"journal-article","created":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T10:34:58Z","timestamp":1765622098000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Frequency-prior enhanced network for facial expression recognition via dynamic large kernels and dual-domain learning"],"prefix":"10.1007","volume":"38","author":[{"given":"Chuanyu","family":"Cai","sequence":"first","affiliation":[]},{"given":"Ke","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,13]]},"reference":[{"key":"414_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2020.165120","volume":"219","author":"W-B An","year":"2020","unstructured":"An W-B, Wang H-M (2020) Infrared and visible image fusion with supervised convolutional neural network. Optik 219:165120. https:\/\/doi.org\/10.1016\/j.ijleo.2020.165120","journal-title":"Optik"},{"issue":"9","key":"414_CR2","doi-asserted-by":"publisher","first-page":"7335","DOI":"10.1016\/j.jksuci.2021.08.021","volume":"34","author":"M Aslan","year":"2022","unstructured":"Aslan M (2022) CNN based efficient approach for emotion recognition. Journal of King Saud University - Computer and Information Sciences 34(9):7335\u20137346","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"key":"414_CR3","doi-asserted-by":"publisher","unstructured":"Chen H, Chu X, Ren Y, Zhao X, Huang K (2024) Pelk: Parameter-efficient large kernel convnets with peripheral convolution. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5557\u20135567. https:\/\/doi.org\/10.1109\/CVPR52733.2024.00531","DOI":"10.1109\/CVPR52733.2024.00531"},{"key":"414_CR4","doi-asserted-by":"publisher","unstructured":"Chen J, Kao Sh, He H, Zhuo W, Wen S, Lee CH, Chan SHG (2023) Run, don\u2019t walk: Chasing higher flops for faster neural networks. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12021\u201312031. https:\/\/doi.org\/10.1109\/CVPR52729.2023.01157","DOI":"10.1109\/CVPR52729.2023.01157"},{"key":"414_CR5","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248\u2013255. Ieee","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"414_CR6","doi-asserted-by":"publisher","unstructured":"Ding, X., Zhang, Y., Ge, Y., Zhao, S., Song, L., Yue, X., Shan, Y.: Unireplknet: A universal perception large-kernel convnet for audio, video, point cloud, time-series and image recognition. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5513\u20135524 (2024). https:\/\/doi.org\/10.1109\/CVPR52733.2024.00527","DOI":"10.1109\/CVPR52733.2024.00527"},{"key":"414_CR7","doi-asserted-by":"publisher","unstructured":"Ding X, Zhang X, Han J, Ding G (2022) Scaling up your kernels to 31$$\\times $$31: Revisiting large kernel design in cnns. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11953\u201311965. https:\/\/doi.org\/10.1109\/CVPR52688.2022.01166","DOI":"10.1109\/CVPR52688.2022.01166"},{"key":"414_CR8","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S et\u00a0al (2020) An image is worth 16x16 words: Transformers for image recognition at scale. arXiv:2010.11929"},{"key":"414_CR9","doi-asserted-by":"crossref","unstructured":"Finder SE, Amoyal R, Treister E, Freifeld O (2024) Wavelet convolutions for large receptive fields. In: European Conference on Computer Vision","DOI":"10.1007\/978-3-031-72949-2_21"},{"key":"414_CR10","doi-asserted-by":"crossref","unstructured":"Goodfellow IJ et\u00a0al (2013) Challenges in representation learning: A report on three machine learning contests. International Conference on Neural Information Processing. Lecture Notes in Computer Science. Springer, Berlin, Germany, pp 117\u2013124","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"414_CR11","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Deep Residual Learning for Image Recognition. arXiv:1512.03385","DOI":"10.1109\/CVPR.2016.90"},{"issue":"2","key":"414_CR12","doi-asserted-by":"publisher","first-page":"639","DOI":"10.1109\/TAFFC.2024.3454102","volume":"16","author":"Y Huang","year":"2025","unstructured":"Huang Y, Peng J, Zhang W, Zhao T, Chen G, Tan S, Yi F, Wang L (2025) FERMixNet: An occlusion robust facial expression recognition model with facial mixing augmentation and mid-level representation learning. IEEE Trans Affect Comput 16(2):639\u2013664. https:\/\/doi.org\/10.1109\/TAFFC.2024.3454102","journal-title":"IEEE Trans Affect Comput"},{"key":"414_CR13","doi-asserted-by":"crossref","unstructured":"Huang G, Liu Z, Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","DOI":"10.1109\/CVPR.2017.243"},{"key":"414_CR14","doi-asserted-by":"crossref","unstructured":"Huang T, Pei X, You S, Wang F, Qian C, Xu C (2024) Localmamba: Visual state space model with windowed selective scan. In: European Conference on Computer Vision, pp. 12\u201322. Springer","DOI":"10.1007\/978-3-031-91979-4_2"},{"key":"414_CR15","unstructured":"Hugo T, Cord M, Matthijs D, Francisco M, Alexandre S, Herv\u00e9 J (2021) Training data-efficient image transformers & distillation through attention. In: ICML"},{"key":"414_CR16","doi-asserted-by":"crossref","unstructured":"Lee-Thorp J, Ainslie J, Eckstein I, Ontanon S (2022) FNet: Mixing Tokens with Fourier Transforms. arXiv:2105.03824","DOI":"10.18653\/v1\/2022.naacl-main.319"},{"key":"414_CR17","unstructured":"Li, W., Cui, Y., Ma, Y., Chen, X., Li, G., Guo, G., Cao, D.: A Spontaneous Driver Emotion Facial Expression (DEFE) Dataset for Intelligent Vehicles (2020). https:\/\/arxiv.org\/abs\/2005.08626"},{"issue":"1","key":"414_CR18","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/TIP.2018.2868382","volume":"28","author":"S Li","year":"2019","unstructured":"Li S, Deng W (2019) Reliable crowdsourcing and deep locality-preserving learning for unconstrained facial expression recognition. IEEE Trans Image Process 28(1):356\u2013370","journal-title":"IEEE Trans Image Process"},{"issue":"1","key":"414_CR19","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/TIP.2018.2868382","volume":"28","author":"S Li","year":"2019","unstructured":"Li S, Deng W (2019) Reliable crowdsourcing and deep locality-preserving learning for unconstrained facial expression recognition. IEEE Trans Image Process 28(1):356\u2013370","journal-title":"IEEE Trans Image Process"},{"issue":"3","key":"414_CR20","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li S, Deng W (2020) Deep facial expression recognition: A survey. IEEE Trans Affect Comput 13(3):1195\u20131215","journal-title":"IEEE Trans Affect Comput"},{"key":"414_CR21","doi-asserted-by":"publisher","unstructured":"Li Y, Lu G, Li J, Zhang Z, Zhang D (2020) Facial expression recognition in the wild using multi-level features and attention mechanisms. IEEE Trans Affective Comput. https:\/\/doi.org\/10.1109\/TAFFC.2020.3031602. To be published","DOI":"10.1109\/TAFFC.2020.3031602"},{"issue":"2","key":"414_CR22","doi-asserted-by":"publisher","first-page":"882","DOI":"10.1109\/TCSVT.2023.3237006","volume":"34","author":"C Li","year":"2024","unstructured":"Li C, Li X, Wang X, Huang D, Liu Z, Liao L (2024) Fg-agr: Fine-grained associative graph representation for facial expression recognition in the wild. IEEE Trans Circuits Syst Video Technol 34(2):882\u2013896. https:\/\/doi.org\/10.1109\/TCSVT.2023.3237006","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"414_CR23","doi-asserted-by":"publisher","unstructured":"Li Y, Hou Q, Zheng Z, Cheng MM, Yang J, Li X (2023) Large selective kernel network for remote sensing object detection. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 16748\u201316759. https:\/\/doi.org\/10.1109\/ICCV51070.2023.01540","DOI":"10.1109\/ICCV51070.2023.01540"},{"key":"414_CR24","doi-asserted-by":"publisher","unstructured":"Liu KH, Shih XK, Liu TJ (2023) Facial expression recognition in the wild from attention to vision transformer based cnns. In: 2023 International Conference on Consumer Electronics - Taiwan (ICCE-Taiwan), pp. 837\u2013838. https:\/\/doi.org\/10.1109\/ICCE-Taiwan58799.2023.10227061","DOI":"10.1109\/ICCE-Taiwan58799.2023.10227061"},{"key":"414_CR25","first-page":"103031","volume":"37","author":"Y Liu","year":"2024","unstructured":"Liu Y, Tian Y, Zhao Y, Yu H, Xie L, Wang Y, Ye Q, Jiao J, Liu Y (2024) Vmamba: Visual state space model. Adv Neural Inf Process Syst 37:103031\u2013103063","journal-title":"Adv Neural Inf Process Syst"},{"key":"414_CR26","unstructured":"Liu S, Chen T, Chen X, Chen X, Xiao Q, Wu B, K\u00e4rkk\u00e4inen T, Pechenizkiy M, Mocanu D, Wang Z (2022) More convnets in the 2020s: Scaling up kernels beyond 51x51 using sparsity. arXiv:2207.03620 (2022)"},{"key":"414_CR27","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"414_CR28","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu CY, Feichtenhofer C, Darrell T, Xie S (2022) A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"414_CR29","doi-asserted-by":"publisher","unstructured":"Li H, Wu XJ, Kittler J (2018) Infrared and visible image fusion using a deep learning framework. In: 2018 24th International Conference on Pattern Recognition (ICPR), pp. 2705\u20132710. https:\/\/doi.org\/10.1109\/ICPR.2018.8546006","DOI":"10.1109\/ICPR.2018.8546006"},{"key":"414_CR30","doi-asserted-by":"publisher","unstructured":"Ma F, Sun B, Li S (2021) Facial expression recognition with visual transformers and attentional selective fusion. IEEE Trans Affect Comput. https:\/\/doi.org\/10.1109\/TAFFC.2021.3122146. Early access, Oct. 26, 2021","DOI":"10.1109\/TAFFC.2021.3122146"},{"issue":"1","key":"414_CR31","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini A, Hasani B, Mahoor MH (2017) Affectnet: A database for facial expression, valence, and arousal computing in the wild. IEEE Trans Affect Comput 10(1):18\u201331","journal-title":"IEEE Trans Affect Comput"},{"issue":"2","key":"414_CR32","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1016\/j.jksuci.2016.12.008","volume":"31","author":"MA Muqeet","year":"2019","unstructured":"Muqeet MA, Holambe RS (2019) Local appearance-based face recognition using adaptive directional wavelet transform. Journal of King Saud University - Computer and Information Sciences 31(2):161\u2013174. https:\/\/doi.org\/10.1016\/j.jksuci.2016.12.008","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"issue":"8","key":"414_CR33","doi-asserted-by":"publisher","first-page":"1030","DOI":"10.1016\/j.jksuci.2019.06.001","volume":"33","author":"R Praditsangthong","year":"2021","unstructured":"Praditsangthong R, Slakkham B, Bhattarakosol P (2021) A fear detection method based on palpebral fissure. Journal of King Saud University - Computer and Information Sciences 33(8):1030\u20131039","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"key":"414_CR34","doi-asserted-by":"publisher","first-page":"428","DOI":"10.1016\/j.procs.2015.03.174","volume":"45","author":"U Ravale","year":"2015","unstructured":"Ravale U, Marathe N, Padiya P (2015) Feature selection based hybrid anomaly intrusion detection system using k means and rbf kernel function. Procedia Computer Science 45:428\u2013435","journal-title":"Procedia Computer Science"},{"issue":"6","key":"414_CR35","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1016\/j.jksuci.2018.09.002","volume":"33","author":"IM Revina","year":"2021","unstructured":"Revina IM, Emmanuel WRS (2021) A survey on human face expression recognition techniques. Journal of King Saud University - Computer and Information Sciences 33(6):619\u2013628. https:\/\/doi.org\/10.1016\/j.jksuci.2018.09.002","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"issue":"3","key":"414_CR36","doi-asserted-by":"publisher","first-page":"392","DOI":"10.1016\/j.jksuci.2018.03.015","volume":"33","author":"IM Revina","year":"2021","unstructured":"Revina IM, Emmanuel WRS (2021) Face expression recognition using LDN and dominant gradient local ternary pattern descriptors. Journal of King Saud University - Computer and Information Sciences 33(3):392\u2013398","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"key":"414_CR37","doi-asserted-by":"publisher","unstructured":"Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520. https:\/\/doi.org\/10.1109\/CVPR.2018.00474","DOI":"10.1109\/CVPR.2018.00474"},{"issue":"2","key":"414_CR38","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","volume":"128","author":"RR Selvaraju","year":"2019","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2019) Grad-cam: Visual explanations from deep networks via gradient-based localization. Int J Comput Vision 128(2):336\u2013359. https:\/\/doi.org\/10.1007\/s11263-019-01228-7","journal-title":"Int J Comput Vision"},{"key":"414_CR39","doi-asserted-by":"crossref","unstructured":"She J, Hu Y, Shi H, Wang J, Shen Q, Mei T (2021) Dive into ambiguity: Latent distribution mining and pairwise uncertainty estimation for facial expression recognition. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., pp. 6248\u20136257","DOI":"10.1109\/CVPR46437.2021.00618"},{"issue":"2","key":"414_CR40","first-page":"176","volume":"9","author":"GS Shergill","year":"2008","unstructured":"Shergill GS, Sarrafzadeh A, Diegel O, Shekar A (2008) Computerized sales assistants: The application of computer technology to measure consumer interest-a conceptual framework. J Electron Commer Res 9(2):176\u2013191","journal-title":"J Electron Commer Res"},{"key":"414_CR41","doi-asserted-by":"crossref","unstructured":"Shi D (2024) Transnext: Robust foveal visual perception for vision transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17773\u201317783","DOI":"10.1109\/CVPR52733.2024.01683"},{"key":"414_CR42","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"414_CR43","unstructured":"Tan M, Le QV (2020) Efficientnet: Rethinking model scaling for convolutional neural networks. arXiv:1905.11946"},{"issue":"2","key":"414_CR44","doi-asserted-by":"publisher","first-page":"258","DOI":"10.1109\/TPAMI.2008.293","volume":"32","author":"Y Tong","year":"2010","unstructured":"Tong Y, Chen J, Ji Q (2010) A unified probabilistic framework for spontaneous facial action modeling and understanding. IEEE Trans Pattern Anal Mach Intell 32(2):258\u2013273. https:\/\/doi.org\/10.1109\/TPAMI.2008.293","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"10","key":"414_CR45","doi-asserted-by":"publisher","first-page":"4109","DOI":"10.1016\/j.jksuci.2020.11.028","volume":"34","author":"M Vasanthi","year":"2022","unstructured":"Vasanthi M, Seetharaman K (2022) Facial image recognition for biometric authentication systems using a combination of geometrical feature points and low-level visual features. Journal of King Saud University - Computer and Information Sciences 34(10):4109\u20134121","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"key":"414_CR46","unstructured":"Wang Y (2018) Face frontalization for facial expression recognition in the wild. PhD thesis, University of Portsmouth"},{"key":"414_CR47","unstructured":"Wang, Y., Yan, S., Liu, Y., Song, W., Liu, J., Chang, Y., Mai, X., Hu, X., Zhang, W., Gan, Z.: A survey on facial expression recognition of static and dynamic emotions. arXiv preprint arXiv:2408.15777 (2024)"},{"key":"414_CR48","doi-asserted-by":"publisher","DOI":"10.1016\/j.jksuci.2023.101605","volume":"35","author":"S Wang","year":"2023","unstructured":"Wang S, Zhao A, Lai C, Zhang Q, Li D, Gao Y, Dong L, Wang X (2023) GCANet: Geometry cues-aware facial expression recognition based on graph convolutional networks. Journal of King Saud University - Computer and Information Sciences 35:101605. https:\/\/doi.org\/10.1016\/j.jksuci.2023.101605","journal-title":"Journal of King Saud University - Computer and Information Sciences"},{"key":"414_CR49","doi-asserted-by":"crossref","unstructured":"Wang K, Peng X, Yang J, Lu S, Qiao Y (2020) Suppressing uncertainties for large-scale facial expression recognition. In: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., pp. 6897\u20136906","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"414_CR50","doi-asserted-by":"publisher","unstructured":"Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Eca-net: Efficient channel attention for deep convolutional neural networks. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11531\u201311539. https:\/\/doi.org\/10.1109\/CVPR42600.2020.01155","DOI":"10.1109\/CVPR42600.2020.01155"},{"issue":"1","key":"414_CR51","doi-asserted-by":"publisher","first-page":"275","DOI":"10.1103\/RevModPhys.78.275","volume":"78","author":"A Wei\u00dfe","year":"2006","unstructured":"Wei\u00dfe A, Wellein G, Alvermann A, Fehske H (2006) The kernel polynomial method. Rev Mod Phys 78(1):275\u2013306","journal-title":"Rev Mod Phys"},{"issue":"6","key":"414_CR52","doi-asserted-by":"publisher","first-page":"668","DOI":"10.1007\/s11263-018-01142-4","volume":"127","author":"Y Wen","year":"2019","unstructured":"Wen Y, Zhang K, Li Z, Qiao Y (2019) A comprehensive study on center loss for deep face recognition. Int J Comput Vision 127(6):668\u2013683","journal-title":"Int J Comput Vision"},{"key":"414_CR53","first-page":"3581","volume-title":"Transfer: Learning relation-aware facial expression representations with transformers","author":"F Xue","year":"2021","unstructured":"Xue F, Wang Q, Guo G (2021) Transfer: Learning relation-aware facial expression representations with transformers. Proc. IEEE\/CVF Int. Conf. Comput. Vis, Montreal, QC, Canada, pp 3581\u20133590"},{"key":"414_CR54","doi-asserted-by":"crossref","unstructured":"Yang J, Qiu P, Zhang Y, Marcus DS, Sotiras A (2024) D-Net: Dynamic Large Kernel with Dynamic Feature Fusion for Volumetric Medical Image Segmentation (2024). arXiv:2403.10674","DOI":"10.2139\/ssrn.5093171"},{"key":"414_CR55","doi-asserted-by":"publisher","first-page":"328","DOI":"10.1007\/978-3-031-19806-9_19","volume-title":"Computer Vision - ECCV 2022","author":"T Yao","year":"2022","unstructured":"Yao T, Pan Y, Li Y, Ngo C-W, Mei T (2022) Wave-vit: Unifying wavelet and transformers for visual representation learning. In: Avidan S, Brostow G, Ciss\u00e9 M, Farinella GM, Hassner T (eds) Computer Vision - ECCV 2022. Springer, Cham, pp 328\u2013345"},{"key":"414_CR56","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2025.127164","volume":"277","author":"J Yu","year":"2025","unstructured":"Yu J, Li S, Tan L, Zhou H, Li Z, Li J (2025) Hivit: Hierarchical attention-based transformer for multi-scale whole slide histopathological image classification. Expert Syst Appl 277:127164","journal-title":"Expert Syst Appl"},{"issue":"6","key":"414_CR57","doi-asserted-by":"publisher","first-page":"2623","DOI":"10.1109\/TIP.2018.2809606","volume":"27","author":"M Zhang","year":"2018","unstructured":"Zhang M, Li W, Du Q (2018) Diverse region-based cnn for hyperspectral image classification. IEEE Trans Image Process 27(6):2623\u20132634","journal-title":"IEEE Trans Image Process"},{"key":"414_CR58","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Liu Y, Sun P, Yan H, Zhao X, Zhang L (2020) Ifcnn: A general image fusion framework based on convolutional neural network. Information Fusion 54:99\u2013118. https:\/\/doi.org\/10.1016\/j.inffus.2019.07.011","journal-title":"Information Fusion"},{"key":"414_CR59","doi-asserted-by":"crossref","unstructured":"Zhang Y, Wang C, Ling X, Deng W (2022) Learn from all: Erasing attention consistency for noisy label facial expression recognition. In: Avidan S, Brostow GJ, Ciss\u00e9 M, Farinella GM, Hassner T. (eds.) Proc. 17th Eur. Conf. Comput. Vis. Lecture Notes in Computer Science, pp. 418\u2013434. Springer, Tel Aviv, Israel","DOI":"10.1007\/978-3-031-19809-0_24"},{"issue":"4","key":"414_CR60","doi-asserted-by":"publisher","first-page":"38","DOI":"10.1109\/MMUL.2021.3107862","volume":"28","author":"X Zhao","year":"2021","unstructured":"Zhao X, Zhu J, Luo B, Gao Y (2021) Survey on facial expression recognition: History, applications, and challenges. IEEE Multimedia 28(4):38\u201344. https:\/\/doi.org\/10.1109\/MMUL.2021.3107862","journal-title":"IEEE Multimedia"},{"key":"414_CR61","unstructured":"Zhao S, Cai H, Liu H, Zhang J, Chen S (2018) Feature selection mechanism in cnns for facial expression recognition. In: Proc. Brit. Mach. Vis. Conf"},{"key":"414_CR62","unstructured":"Zhu L, Liao B, Zhang Q, Wang X, Liu W, Wang X (2024) Vision mamba: Efficient visual representation learning with bidirectional state space model. arXiv:2401.09417"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00414-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-025-00414-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00414-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T14:38:45Z","timestamp":1773153525000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-025-00414-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,13]]},"references-count":62,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["414"],"URL":"https:\/\/doi.org\/10.1007\/s44443-025-00414-7","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,13]]},"assertion":[{"value":"23 September 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable. This study did not use sensitive personal data requiring formal ethics approval.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"The source code used in this study is available from the corresponding author, upon reasonable request.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability"}}],"article-number":"29"}}