{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T15:05:53Z","timestamp":1775228753766,"version":"3.50.1"},"reference-count":56,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,2,2]],"date-time":"2026-02-02T00:00:00Z","timestamp":1769990400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T00:00:00Z","timestamp":1775174400000},"content-version":"vor","delay-in-days":60,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"the National Key Laboratory","award":["SKLK22-11"],"award-info":[{"award-number":["SKLK22-11"]}]},{"name":"the Key Research and Development Program of Shaanxi Province","award":["K20220022"],"award-info":[{"award-number":["K20220022"]}]},{"name":"the Special Project for Local Service of Shaanxi Provincial Department of Education","award":["23JC039"],"award-info":[{"award-number":["23JC039"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s44443-026-00511-1","type":"journal-article","created":{"date-parts":[[2026,2,2]],"date-time":"2026-02-02T21:03:31Z","timestamp":1770066211000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["STAR-Former: spatio-temporal adaptive and region-aware transformer for dynamic facial expression recognition"],"prefix":"10.1007","volume":"38","author":[{"given":"Daipeng","family":"Guo","sequence":"first","affiliation":[]},{"given":"Fei","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,2]]},"reference":[{"key":"511_CR1","unstructured":"Alexey D, Lucas B, Alexander K, Dirk W, Xiaohua Z, Thomas U, Mostafa D, Matthias M, Georg H, Sylvain G, Jakob U, Neil H (2020) An image is worth 16x16 words: transformers for image recognition at scale. In: 2021 International Conference on Learning Representations(ICLR)."},{"key":"511_CR2","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer vision \u2013 ECCV 2020","author":"N Carion","year":"2020","unstructured":"Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-End Object Detection with Transformers. In: Vedaldi A, Bischof H, Brox T, Frahm J-M (eds) Computer vision \u2013 ECCV 2020. Springer International Publishing, Cham, pp 213\u2013229"},{"key":"511_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.123635","volume":"249","author":"D Chen","year":"2024","unstructured":"Chen D, Wen G, Li H, Yang P, Chen C, Wang B (2024) Multi-geometry embedded transformer for facial expression recognition in videos. Expert Syst Appl 249:123635","journal-title":"Expert Syst Appl"},{"key":"511_CR4","unstructured":"Chung J, G\u00fcl\u00e7ehre \u00c7, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling, CoRR abs\/1412.3555"},{"key":"511_CR5","doi-asserted-by":"publisher","first-page":"160","DOI":"10.1016\/S1077-3142(03)00081-X","volume":"91","author":"I Cohen","year":"2003","unstructured":"Cohen I, Sebe N, Garg A, Chen LS, Huang TS (2003) Facial expression recognition from video sequences: temporal and static modeling. Comput vis Image Underst 91:160\u2013187","journal-title":"Comput vis Image Underst"},{"key":"511_CR6","doi-asserted-by":"publisher","DOI":"10.1093\/oso\/9780195392289.001.0001","volume-title":"The expression of the emotions in man and animals","author":"C Darwin","year":"2009","unstructured":"Darwin C (2009) The expression of the emotions in man and animals. Cambridge University Press, Cambridge"},{"key":"511_CR7","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1109\/MMUL.2012.26","volume":"19","author":"A Dhall","year":"2012","unstructured":"Dhall A, Goecke R, Lucey S, Gedeon T (2012) Collecting large, richly annotated facial-expression databases from movies. IEEE Multimedia 19:34\u201341","journal-title":"IEEE Multimedia"},{"key":"511_CR8","doi-asserted-by":"crossref","unstructured":"Dhall A, Goecke R, Joshi J, Wagner M, Gedeon T (2013) Emotion recognition in the wild challenge 2013. In: Proceedings of the 15th ACM on International conference on multimodal interaction, Association for Computing Machinery, Sydney, Australia, pp. 509\u2013516","DOI":"10.1145\/2522848.2531739"},{"key":"511_CR9","doi-asserted-by":"crossref","unstructured":"Fan Y, Lu X, Li D, Liu Y (2016) Video-based emotion recognition using CNN-RNN and C3D hybrid networks. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction, Association for Computing Machinery, Tokyo, Japan, pp. 445\u2013450","DOI":"10.1145\/2993148.2997632"},{"key":"511_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2024.109125","volume":"115","author":"Z Han","year":"2024","unstructured":"Han Z, Meichen X, Hong P, Zhicai L, Jun G (2024) Nsnp-dfer: a nonlinear spiking neural p network for dynamic facial expression recognition. Comput Electr Eng 115:109125","journal-title":"Comput Electr Eng"},{"key":"511_CR11","doi-asserted-by":"crossref","unstructured":"Hara K, Kataoka H, Satoh Y (2018) Can spatiotemporal 3D CNNs retrace the history of 2D CNNs and ImageNet?. In: 2018 IEEE\/CVF conference on computer vision and pattern recognition, pp. 6546\u20136555","DOI":"10.1109\/CVPR.2018.00685"},{"key":"511_CR12","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"511_CR13","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Long Short-Term Memory. Neural Comput 9:1735\u20131780","journal-title":"Neural Comput"},{"key":"511_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2023.107480","volume":"233","author":"C-W Huang","year":"2023","unstructured":"Huang C-W, Wu BCY, Nguyen PA, Wang H-H, Kao C-C, Lee P-C, Rahmanti AR, Hsu JC, Yang H-C, Li Y-CJ (2023) Emotion recognition in doctor-patient interactions from real-world clinical video database: initial development of artificial empathy. Comput Methods Programs Biomed 233:107480","journal-title":"Comput Methods Programs Biomed"},{"key":"511_CR15","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2024.108535","volume":"133","author":"Z Huang","year":"2024","unstructured":"Huang Z, Zhu Y, Li H, Yang D (2024) Dynamic facial expression recognition based on spatial key-points optimized region feature fusion and temporal self-attention. Eng Appl Artif Intell 133:108535","journal-title":"Eng Appl Artif Intell"},{"key":"511_CR16","doi-asserted-by":"publisher","DOI":"10.3390\/s18124270","author":"M Jeong","year":"2018","unstructured":"Jeong M, Ko BC (2018) Driver\u2019s facial expression recognition in real-time for safe driving. Sensors. https:\/\/doi.org\/10.3390\/s18124270","journal-title":"Sensors"},{"key":"511_CR17","doi-asserted-by":"crossref","unstructured":"Jiang X, Zong Y, Zheng W, Tang C, Xia W, Lu C, Liu J (2020) DFEW: a large-scale database for recognizing dynamic facial expressions in the wild. In: Proceedings of the 28th ACM international conference on multimedia, association for computing machinery, Seattle, WA, USA, pp. 2881\u20132889","DOI":"10.1145\/3394171.3413620"},{"key":"511_CR18","doi-asserted-by":"crossref","unstructured":"Kahou SE, Michalski V, Konda K, Memisevic R, Pal C (2015) Recurrent neural networks for emotion recognition in video. In: Proceedings of the 2015 ACM on international conference on multimodal interaction, association for computing machinery, Seattle, Washington, USA, pp. 467\u2013474","DOI":"10.1145\/2818346.2830596"},{"key":"511_CR19","doi-asserted-by":"publisher","first-page":"223","DOI":"10.1109\/TAFFC.2017.2695999","volume":"10","author":"DH Kim","year":"2019","unstructured":"Kim DH, Baddar WJ, Jang J, Ro YM (2019) Multi-objective based spatio-temporal feature representation learning robust to expression intensity variations for facial expression recognition. IEEE Trans Affect Comput 10:223\u2013236","journal-title":"IEEE Trans Affect Comput"},{"key":"511_CR20","doi-asserted-by":"crossref","unstructured":"Kossaifi J, Toisoul A, Bulat A, Panagakis Y, Hospedales TM, Pantic M (2020) Factorized higher-order CNNs with an application to spatio-temporal emotion estimation. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6059\u20136068","DOI":"10.1109\/CVPR42600.2020.00610"},{"key":"511_CR21","unstructured":"Kunyu P, Di W, Jianyang F, Jiamin W, Kailun Y, Junwei Z, Ruiping L, Yufan C, Yuqian F, Danda Pani P, Luc van G, Rainer S (2025) RefAtomNet++: Advancing referring atomic video action recognition using semantic retrieval based multi-trajectory mamba, ArXiv abs\/2510.16444"},{"key":"511_CR22","doi-asserted-by":"crossref","unstructured":"Lee B, Shin H, Ku B, Ko H (2023) Frame level emotion guided dynamic facial expression recognition with emotion grouping. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 5681\u20135691","DOI":"10.1109\/CVPRW59228.2023.00602"},{"key":"511_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2024.120953","volume":"678","author":"M Li","year":"2024","unstructured":"Li M, Zhang X, Fan C, Liao T, Xiao G (2024) Dual-STI: Dual-path spatial-temporal interaction learning for dynamic facial expression recognition. Inf Sci 678:120953","journal-title":"Inf Sci"},{"key":"511_CR24","doi-asserted-by":"crossref","unstructured":"Li Q, Liu X, Gong X, Jing S (2019) INDReview on Facial expression analysis and its application in education. In: 2019 Chinese Automation Congress (CAC), pp. 4526\u20134530.","DOI":"10.1109\/CAC48633.2019.8996796"},{"key":"511_CR25","unstructured":"Li H, Sui M, Zhu Z, Zhao F (2022) NR-DFERNet: noise-robust network for dynamic facial expression recognition, CoRR abs\/2206.04975"},{"key":"511_CR26","doi-asserted-by":"crossref","unstructured":"Li H, Niu H, Zhu Z, Zhao F (2023) Intensity-aware loss for dynamic facial expression recognition in the wild. In: Proceedings of the thirty-seventh AAAI conference on artificial intelligence and thirty-fifth conference on innovative applications of artificial intelligence and thirteenth symposium on educational advances in artificial intelligence, AAAI Press, p. Article 8","DOI":"10.1609\/aaai.v37i1.25077"},{"key":"511_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108105","volume":"119","author":"X Liu","year":"2021","unstructured":"Liu X, Jin L, Han X, You J (2021a) Mutual information regularized identity-aware facial expression recognition in compressed video. Pattern Recognit 119:108105","journal-title":"Pattern Recognit"},{"key":"511_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109368","volume":"138","author":"Y Liu","year":"2023","unstructured":"Liu Y, Wang W, Feng C, Zhang H, Chen Z, Zhan Y (2023) Expression snippet transformer for robust video-based facial expression recognition. Pattern Recogn 138:109368","journal-title":"Pattern Recogn"},{"key":"511_CR29","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021b) Swin transformer: hierarchical vision transformer using shifted windows. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9992\u201310002","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"511_CR30","doi-asserted-by":"crossref","unstructured":"Liu F, Wang H, Shen S (2025) Robust dynamic facial expression recognition. IEEE transactions on biometrics, behavior, and identity science, 1\u20131","DOI":"10.1109\/TBIOM.2025.3546279"},{"key":"511_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2024.104915","volume":"142","author":"C Lu","year":"2024","unstructured":"Lu C, Jiang Y, Fu K, Zhao Q, Yang H (2024) LSTPNet: Long short-term perception network for dynamic facial expression recognition in the wild. Image vis Comput 142:104915","journal-title":"Image vis Comput"},{"key":"511_CR32","doi-asserted-by":"crossref","unstructured":"Lu C, Zheng W, Li C, Tang C, Liu S, Yan S, Zong Y (2018) Multiple spatio-temporal feature learning for video-based emotion recognition in the wild. In: Proceedings of the 20th ACM international conference on multimodal interaction, association for computing machinery, Boulder, CO, USA, pp. 646\u2013652","DOI":"10.1145\/3242969.3264992"},{"key":"511_CR33","doi-asserted-by":"crossref","unstructured":"Ma F, Sun B, Li S (2023) Logo-former: local-global spatio-temporal transformer for dynamic facial expression recognition. In: ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1\u20135","DOI":"10.1109\/ICASSP49357.2023.10095448"},{"key":"511_CR34","doi-asserted-by":"crossref","unstructured":"Meng D, Peng X, Wang K, Qiao Y (2019) Frame attention networks for facial expression recognition in videos. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 3866\u20133870.","DOI":"10.1109\/ICIP.2019.8803603"},{"key":"511_CR35","doi-asserted-by":"publisher","DOI":"10.3390\/electronics13091606","author":"N Neshov","year":"2024","unstructured":"Neshov N, Christoff N, Sechkova T, Tonchev K, Manolova A (2024) SlowR50-SA: a self-attention enhanced dynamic facial expression recognition model for tactile internet applications. Electronics. https:\/\/doi.org\/10.3390\/electronics13091606","journal-title":"Electronics"},{"key":"511_CR36","doi-asserted-by":"crossref","unstructured":"Peng K, Fu J, Yang K, Wen D, Chen Y, Liu R, Zheng J, Zhang J, Sarfraz MS, Stiefelhagen R, Roitberg A (2025) Referring atomic video action recognition. In: Leonardis A, Ricci E, Roth S, Russakovsky O, Sattler T, Varol G (eds) Computer vision \u2013 ECCV 2024, Springer Nature Switzerland, Cham, pp. 166\u2013185","DOI":"10.1007\/978-3-031-72655-2_10"},{"key":"511_CR37","doi-asserted-by":"crossref","unstructured":"Sikka K, Dykstra K, Sathyanarayana S, Littlewort G, Bartlett M (2013) Multiple kernel learning for emotion recognition in the wild. In: Proceedings of the 15th ACM on International conference on multimodal interaction, Association for Computing Machinery, Sydney, Australia, pp. 517\u2013524","DOI":"10.1145\/2522848.2531741"},{"key":"511_CR38","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition, in: 3rd International Conference on Learning Representations (ICLR 2015), Computational and Biological Learning Society, pp. 1\u201314"},{"key":"511_CR39","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015) Learning spatiotemporal features with 3D convolutional networks. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 4489\u20134497","DOI":"10.1109\/ICCV.2015.510"},{"key":"511_CR40","first-page":"2579","volume":"9","author":"L van der Maaten","year":"2008","unstructured":"van der Maaten L, Hinton G (2008) Viualizing data using t-SNE. J Mach Learn Res 9:2579\u20132605","journal-title":"J Mach Learn Res"},{"key":"511_CR41","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, Curran Associates Inc., Long Beach, California, USA, pp. 6000\u20136010"},{"key":"511_CR42","doi-asserted-by":"crossref","unstructured":"Wang R, Sun X (2023) Dynamic facial expression recognition based on vision transformer with deformable module. In: 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 2138\u20132143","DOI":"10.1109\/SMC53992.2023.10394270"},{"key":"511_CR43","doi-asserted-by":"crossref","unstructured":"Wang Y, Sun Y, Huang Y, Liu Z, Gao S, Zhang W, Ge W, Zhang W (2022a) FERV39k: A large-scale multi-scene dataset for facial expression recognition in videos. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 20890\u201320899","DOI":"10.1109\/CVPR52688.2022.02025"},{"key":"511_CR44","doi-asserted-by":"crossref","unstructured":"Wang Y, Sun Y, Song W, Gao S, Huang Y, Chen Z, Ge W, Zhang W (2022b) DPCNet: Dual path multi-excitation collaborative network for facial expression representation learning in videos. In: Proceedings of the 30th ACM international conference on multimedia, association for computing machinery, Lisboa, Portugal, pp. 101\u2013110","DOI":"10.1145\/3503161.3547865"},{"key":"511_CR45","doi-asserted-by":"crossref","unstructured":"Wang H, Li B, Wu S, Shen S, Liu F, Ding S, Zhou A (2023) Rethinking the learning paradigm for dynamic facial expression recognition. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 17958\u201317968","DOI":"10.1109\/CVPR52729.2023.01722"},{"key":"511_CR46","doi-asserted-by":"crossref","unstructured":"Wang H, Mai X, Tao Z, Tong X, Lin J, Wang Y, Yu J, Wang B, Yan S, Zhao Q, Zhou Z, Gao S, Zhang W (2024a) D2SP: Dynamic dual-stage purification framework for dual noise mitigation in vision-based affective recognition in 2024.","DOI":"10.1109\/CVPR52734.2025.01790"},{"key":"511_CR47","doi-asserted-by":"crossref","unstructured":"Wang L, Kang X, Ding F, Nakagawa S, Ren F (2024b) MSSTNet: A multi-scale spatio-temporal CNN-transformer network for dynamic facial expression recognition. In: ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3015\u20133019","DOI":"10.1109\/ICASSP48485.2024.10446699"},{"key":"511_CR48","doi-asserted-by":"publisher","first-page":"130","DOI":"10.1016\/j.comcom.2023.12.032","volume":"216","author":"S Yan","year":"2024","unstructured":"Yan S, Wang Y, Mai X, Zhao Q, Song W, Huang J, Tao Z, Wang H, Gao S, Zhang W (2024) Empower smart cities with sampling-wise dynamic facial expression recognition via frame-sequence contrastive learning. Comput Commun 216:130\u2013139","journal-title":"Comput Commun"},{"key":"511_CR49","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2024.121415","volume":"689","author":"S Yan","year":"2025","unstructured":"Yan S, Wang Y, Mai X, Tao Z, Song W, Zhao Q, Wang B, Wang H, Gao S, Zhang W (2025) Observe finer to select better: learning key frame extraction via semantic coherence for dynamic facial expression recognition in the wild. Inf Sci 689:121415","journal-title":"Inf Sci"},{"key":"511_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108401","volume":"123","author":"W Yu","year":"2022","unstructured":"Yu W, Xu H (2022) Co-attentive multi-task convolutional neural network for facial expression recognition. Pattern Recognit 123:108401","journal-title":"Pattern Recognit"},{"key":"511_CR51","doi-asserted-by":"publisher","first-page":"10503","DOI":"10.1109\/TMM.2024.3407693","volume":"26","author":"Z Zhang","year":"2024","unstructured":"Zhang Z, Tian X, Zhang Y, Guo K, Xu X (2024a) Label-guided dynamic spatial-temporal fusion for video-based facial expression recognition. IEEE Trans Multimedia 26:10503\u201310513","journal-title":"IEEE Trans Multimedia"},{"key":"511_CR52","doi-asserted-by":"publisher","first-page":"3192","DOI":"10.1109\/TCSVT.2023.3312858","volume":"34","author":"X Zhang","year":"2024","unstructured":"Zhang X, Li M, Lin S, Xu H, Xiao G (2024b) Transformer-based multimodal emotional perception for dynamic facial expression recognition in the wild. IEEE Trans Circuits Syst Video Technol 34:3192\u20133203","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"511_CR53","doi-asserted-by":"crossref","unstructured":"Zhang YH, Huang R, Zeng J, Shan S (2020) M3F: multi-modal continuous valence-arousal estimation in the wild. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), pp. 632\u2013636","DOI":"10.1109\/FG47880.2020.00098"},{"key":"511_CR54","doi-asserted-by":"publisher","first-page":"2751","DOI":"10.1109\/TAFFC.2022.3181736","volume":"14","author":"R Zhao","year":"2023","unstructured":"Zhao R, Liu T, Huang Z, Lun DPK, Lam KM (2023) Spatial-temporal graphs plus transformers for geometry-guided facial expression recognition. IEEE Trans Affect Comput 14:2751\u20132767","journal-title":"IEEE Trans Affect Comput"},{"key":"511_CR55","doi-asserted-by":"crossref","unstructured":"Zhao Z, Liu Q (2021) Former-DFER: dynamic facial expression recognition transformer. In: Proceedings of the 29th ACM international conference on multimedia, association for computing machinery, virtual event, China, pp. 1553\u20131561","DOI":"10.1145\/3474085.3475292"},{"key":"511_CR56","doi-asserted-by":"crossref","unstructured":"Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) Learning deep features for discriminative localization. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2921\u20132929","DOI":"10.1109\/CVPR.2016.319"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-026-00511-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00511-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00511-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T14:18:57Z","timestamp":1775225937000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-026-00511-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,2]]},"references-count":56,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["511"],"URL":"https:\/\/doi.org\/10.1007\/s44443-026-00511-1","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,2]]},"assertion":[{"value":"11 November 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"111"}}