{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T09:11:33Z","timestamp":1773220293164,"version":"3.50.1"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,1,24]],"date-time":"2026-01-24T00:00:00Z","timestamp":1769212800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T00:00:00Z","timestamp":1773100800000},"content-version":"vor","delay-in-days":45,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s44443-025-00447-y","type":"journal-article","created":{"date-parts":[[2026,1,24]],"date-time":"2026-01-24T10:54:18Z","timestamp":1769252058000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing multimodal sentiment analysis reliability: SentiGuard+ with Dirichlet evidence and selective prediction"],"prefix":"10.1007","volume":"38","author":[{"given":"Komal Rani","family":"Narejo","sequence":"first","affiliation":[]},{"given":"Hongying","family":"Zan","sequence":"additional","affiliation":[]},{"given":"Samih M.","family":"Mostafa","sequence":"additional","affiliation":[]},{"given":"Faten Khalid","family":"Karim","sequence":"additional","affiliation":[]},{"given":"Faisal","family":"Mehmood","sequence":"additional","affiliation":[]},{"given":"Ayesha","family":"Yaseen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,24]]},"reference":[{"issue":"6","key":"447_CR1","doi-asserted-by":"publisher","first-page":"103857","DOI":"10.1016\/j.ipm.2024.103857","volume":"61","author":"N B\u00f6l\u00fcc\u00fc","year":"2024","unstructured":"B\u00f6l\u00fcc\u00fc N, Rybinski M, Dai X, Wan S (2024) An adaptive approach to noisy annotations in scientific information extraction. Inform Process Manag 61(6):103857","journal-title":"Inform Process Manag"},{"issue":"1","key":"447_CR2","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1109\/TIT.1970.1054413","volume":"16","author":"CK Chow","year":"1970","unstructured":"Chow CK (1970) On optimum recognition error and reject tradeoff. IEEE Trans Inf Theory 16(1):41\u201346. https:\/\/doi.org\/10.1109\/TIT.1970.1054413","journal-title":"IEEE Trans Inf Theory"},{"key":"447_CR3","doi-asserted-by":"crossref","unstructured":"Cortes C, DeSalvo G, Mohri M (2016) Learning with rejection. In: Proceedings of Algorithmic Learning Theory (ALT)","DOI":"10.1007\/978-3-319-46379-7_5"},{"issue":"13s","key":"447_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3586075","volume":"55","author":"R Das","year":"2023","unstructured":"Das R, Singh TD (2023) Multimodal sentiment analysis: a survey of methods, trends, and challenges. ACM Comput Surv 55(13s):1\u201338","journal-title":"ACM Comput Surv"},{"key":"447_CR5","unstructured":"DeVries T, Taylor GW (2018) Learning confidence for out-of-distribution detection in neural networks. arXiv:1802.04865"},{"key":"447_CR6","doi-asserted-by":"publisher","first-page":"101955","DOI":"10.1016\/j.inffus.2023.101955","volume":"100","author":"D Folgado","year":"2023","unstructured":"Folgado D, Barandas M, Famiglini L, Santos R, Cabitza F, Gamboa H (2023) Explainability meets uncertainty quantification: Insights from feature-based model fusion on multimodal time series. Information Fusion 100:101955. https:\/\/doi.org\/10.1016\/j.inffus.2023.101955","journal-title":"Information Fusion"},{"key":"447_CR7","unstructured":"Gal Y, Ghahramani Z (2016) Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In: ICML. pp 1050\u20131059"},{"issue":"5","key":"447_CR8","doi-asserted-by":"publisher","first-page":"829","DOI":"10.1162\/neco_a_01273","volume":"32","author":"J Gao","year":"2020","unstructured":"Gao J, Li P, Chen Z, Zhang J (2020) A survey on deep learning for multimodal data fusion. Neural Comput 32(5):829\u2013864. https:\/\/doi.org\/10.1162\/neco_a_01273","journal-title":"Neural Comput"},{"key":"447_CR9","unstructured":"Geifman Y, El-Yaniv R (2017) Selective classification for deep neural networks. In: Proceedings of NeurIPS"},{"key":"447_CR10","unstructured":"Geifman Y, Uziel G, El-Yaniv R (2019) Selectivenet: A deep neural network with an integrated reject option. In: Proceedings of ICML. https:\/\/proceedings.mlr.press\/v97\/geifman19a.html"},{"key":"447_CR11","unstructured":"Guo C, Pleiss G, Sun Y, Weinberger KQ (2017) On calibration of modern neural networks. In: Proceedings of ICML. https:\/\/proceedings.mlr.press\/v70\/guo17a.html"},{"key":"447_CR12","doi-asserted-by":"publisher","unstructured":"Hazarika D, Zimmermann R, Poria S (2020) Misa: Modality-invariant and -specific representations for multimodal sentiment analysis. In: Proceedings of the 28th ACM International Conference on Multimedia (MM \u201920). https:\/\/doi.org\/10.1145\/3394171.3413678. https:\/\/dl.acm.org\/doi\/10.1145\/3394171.3413678","DOI":"10.1145\/3394171.3413678"},{"key":"447_CR13","unstructured":"Hendrycks D, Gimpel K (2017) A baseline for detecting misclassified and out-of-distribution examples in neural networks. In: ICLR"},{"key":"447_CR14","unstructured":"Hu, M., Zhang, Z., Zhao, S., Huang, M., Wu, B. (2023) Uncertainty in natural language processing: Sources, quantification, and applications. arXiv:2306.04459"},{"key":"447_CR15","doi-asserted-by":"crossref","unstructured":"Karun SP, Adithya V (2025) Applying cross-modal feature alignment and fusion for effective sarcasm detection. Progr Artif Intell 1\u201313","DOI":"10.1007\/s13748-025-00370-3"},{"key":"447_CR16","unstructured":"Kose N, Krishnan R, Dhamasia A, Tickoo O, Paulitsch M (2022) Reliable uncertainty via error-aligned loss. https:\/\/arxiv.org\/abs\/2207.11893"},{"key":"447_CR17","unstructured":"Kull M, Perell\u00f3-Nieto M, K\u00e4ngsepp M, Silva\u00a0Filho TdMe, Song H, Flach P (2019) Beyond temperature scaling: Obtaining well-calibrated multi-class probabilities with dirichlet calibration. In: Advances in Neural Information Processing Systems, vol 32"},{"issue":"17","key":"447_CR18","doi-asserted-by":"publisher","first-page":"24103","DOI":"10.1007\/s11042-019-7390-1","volume":"78","author":"A Kumar","year":"2019","unstructured":"Kumar A, Garg G (2019) Sentiment analysis of multimodal twitter data. Multim Tools Appl 78(17):24103\u201324119","journal-title":"Multim Tools Appl"},{"key":"447_CR19","unstructured":"Lakshminarayanan B, Pritzel A, Blundell C (2017) Simple and scalable predictive uncertainty estimation using deep ensembles. In: Proceedings of NeurIPS. pp 6402\u20136413. https:\/\/papers.nips.cc\/paper\/2017\/hash\/9ef2ed4b7fd2c810847ffa5fa85bce38-Abstract.html"},{"issue":"11","key":"447_CR20","doi-asserted-by":"publisher","first-page":"2069","DOI":"10.3390\/electronics13112069","volume":"13","author":"X Li","year":"2024","unstructured":"Li X et al (2024) Multi-modal sentiment analysis based on image and text fusion via cross-attention (mcam). Electronics 13(11):2069. https:\/\/doi.org\/10.3390\/electronics13112069","journal-title":"Electronics"},{"key":"447_CR21","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1007\/s44443-025-00094-3","volume":"37","author":"X Li","year":"2025","unstructured":"Li X, Ma Y, An X et al (2025) Multi-level fusion with fine-grained alignment for multimodal sentiment analysis. J King Saud Univ Comput Inf Sci 37:82. https:\/\/doi.org\/10.1007\/s44443-025-00094-3","journal-title":"J King Saud Univ Comput Inf Sci"},{"issue":"10","key":"447_CR22","doi-asserted-by":"publisher","first-page":"11184","DOI":"10.1007\/s10489-021-02936-9","volume":"52","author":"W Liao","year":"2022","unstructured":"Liao W, Zeng B, Liu J, Wei P, Fang J (2022) Image-text interaction graph neural network for image-text sentiment analysis. Appl Intell 52(10):11184\u201311198","journal-title":"Appl Intell"},{"key":"447_CR23","doi-asserted-by":"publisher","first-page":"102353","DOI":"10.1016\/j.inffus.2024.102353","volume":"108","author":"H Liu","year":"2024","unstructured":"Liu H, Wei R, Tu G, Lin J, Liu C, Jiang D (2024) Sarcasm driven by sentiment: a sentiment-aware hierarchical fusion network for multimodal sarcasm detection. Inform Fusion 108:102353","journal-title":"Inform Fusion"},{"issue":"2","key":"447_CR24","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/s11063-025-11737-x","volume":"57","author":"Z Liu","year":"2025","unstructured":"Liu Z, Lin J, Chen Y, Dong Y (2025) Multimodal aspect-based sentiment analysis with external knowledge and multi-granularity image-text features. Neural Process Lett 57(2):25","journal-title":"Neural Process Lett"},{"key":"447_CR25","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV). IEEE, pp. 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"447_CR26","unstructured":"Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) Roberta: a robustly optimized Bert pretraining approach. arXiv:1907.11692"},{"key":"447_CR27","unstructured":"Liu Z, Wang Z, Liang PP, Salakhutdinov R, Morency L, Ueda M (2019) Deep gamblers: Learning to abstain with portfolio theory. In: Proceedings of NeurIPS"},{"key":"447_CR28","doi-asserted-by":"publisher","unstructured":"Lu D, Neves L, Carvalho V, Zhang N, Ji H (2018) Visual attention model for name tagging in multimodal social media. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL). pp 1990\u20131999.https:\/\/doi.org\/10.18653\/v1\/P18-1185","DOI":"10.18653\/v1\/P18-1185"},{"key":"447_CR29","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1016\/j.knosys.2018.07.040","volume":"161","author":"N Majumder","year":"2018","unstructured":"Majumder N, Hazarika D, Gelbukh A, Cambria E, Poria S (2018) Multimodal sentiment analysis using hierarchical fusion with context modeling. Knowl-Based Syst 161:124\u2013133. https:\/\/doi.org\/10.1016\/j.knosys.2018.07.040","journal-title":"Knowl-Based Syst"},{"key":"447_CR30","unstructured":"Malinin A, Gales M (2018) Predictive uncertainty estimation via prior networks. arXiv:1802.10501"},{"key":"447_CR31","unstructured":"Ma H, Zhang Q, Zhang C, Wu B, Fu H, Zhou JT, Hu Q (2023) Calibrating multimodal learning. In: Proceedings of the 40th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol 202. PMLR, Honolulu, Hawaii, USA, pp 23429\u201323450. https:\/\/proceedings.mlr.press\/v202\/ma23i.html"},{"key":"447_CR32","doi-asserted-by":"publisher","unstructured":"Niu T, Zhu S, Pang L, El\u00a0Saddik A (2016) Sentiment analysis on multi-view social data. In: MultiMedia Modeling (MMM). Miami, USA, pp 15\u201327. https:\/\/doi.org\/10.1007\/978-3-319-27674-8_2","DOI":"10.1007\/978-3-319-27674-8_2"},{"key":"447_CR33","doi-asserted-by":"publisher","first-page":"128874","DOI":"10.1016\/j.neucom.2024.128874","volume":"616","author":"L Ou","year":"2025","unstructured":"Ou L, Li Z (2025) Modeling inter-modal incongruous sentiment expressions for multi-modal sarcasm detection. Neurocomputing 616:128874","journal-title":"Neurocomputing"},{"issue":"3","key":"447_CR34","doi-asserted-by":"publisher","first-page":"662","DOI":"10.3390\/electronics13030662","volume":"13","author":"Q Pan","year":"2024","unstructured":"Pan Q, Meng Z (2024) Hybrid uncertainty calibration for multimodal sentiment analysis. Electronics 13(3):662","journal-title":"Electronics"},{"issue":"18","key":"447_CR35","first-page":"54249","volume":"83","author":"A Paul","year":"2024","unstructured":"Paul A, Nayyar A (2024) A context-sensitive multi-tier deep learning framework for multimodal sentiment analysis. Multim Tools Appl 83(18):54249\u201354278","journal-title":"Multim Tools Appl"},{"issue":"10","key":"447_CR36","doi-asserted-by":"publisher","first-page":"1922","DOI":"10.3390\/electronics13101922","volume":"13","author":"K Qiu","year":"2024","unstructured":"Qiu K et al (2024) Multimodal sentiment analysis with variational autoencoder and joint chained interactive attention (vae-jcia). Electronics 13(10):1922. https:\/\/doi.org\/10.3390\/electronics13101922","journal-title":"Electronics"},{"key":"447_CR37","unstructured":"Qiu J, Zhu Y, Shi X, Wenzel F, Tang Z, Zhao D, Li B, Li M (2022) Benchmarking robustness of multimodal image-text models under distribution shift. arXiv:2212.08044"},{"key":"447_CR38","doi-asserted-by":"publisher","unstructured":"Rahman WU, Hasan MK, Zadeh AB, Morency L-P (2020) Integrating multimodal adaptation gate in pretrained transformers for multimodal sentiment analysis. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp 2359\u20132369. Association for Computational Linguistics). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.214. https:\/\/aclanthology.org\/2020.acl-main.214","DOI":"10.18653\/v1\/2020.acl-main.214"},{"key":"447_CR39","doi-asserted-by":"crossref","unstructured":"Satthar FS, Evans R, Uchyigit G (2017) A calibration method for evaluation of sentiment analysis. In: Proceedings of RANLP. pp 652\u2013660","DOI":"10.26615\/978-954-452-049-6_084"},{"key":"447_CR40","doi-asserted-by":"crossref","unstructured":"Schuff H, Barnes J, Mohme J, Pad\u00f3 S, Klinger R (2017) Annotation, modelling and analysis of fine-grained emotions on a stance and sentiment detection corpus. In: Proceedings of the 8th workshop on computational approaches to subjectivity, sentiment and social media analysis. Association for Computational Linguistics, Copenhagen, Denmark, pp 13\u201323","DOI":"10.18653\/v1\/W17-5203"},{"key":"447_CR41","unstructured":"Sensoy M, Kaplan L, Kandemir M (2018) Evidential deep learning to quantify classification uncertainty. In: Proceedings of NeurIPS"},{"key":"447_CR42","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-025-03883-z","author":"D Shao","year":"2025","unstructured":"Shao D, Wang J, Li Z et al (2025) Transformer-based short-term memory attention for enhanced multimodal sentiment analysis. Vis Comput. https:\/\/doi.org\/10.1007\/s00371-025-03883-z","journal-title":"Vis Comput"},{"key":"447_CR43","doi-asserted-by":"crossref","unstructured":"Shao D, Tang K, Li J, Yi S, Ma L (2025) Transformer-based short-term memory attention for enhanced multimodal sentiment analysis. Vis Comput 1\u201316","DOI":"10.1007\/s00371-025-03883-z"},{"key":"447_CR44","unstructured":"Tagasovska N, Lopez-Paz D (2019) Single-model uncertainties for deep learning. In: NeurIPS"},{"key":"447_CR45","doi-asserted-by":"publisher","unstructured":"Tan X, Gong Z, Gan M et al (2025) Graph convolutional network model with a feature compensation module and dual-channel second-order pooling module for multimodal emotion recognition in conversation. J King Saud Univ Comput Inf Sci 37:93. https:\/\/doi.org\/10.1007\/s44443-025-00091-6","DOI":"10.1007\/s44443-025-00091-6"},{"key":"447_CR46","doi-asserted-by":"crossref","unstructured":"Tellamekala MK, Amiriparian S, Schuller BW, Andr\u00e9 E, Giesbrecht T, Valstar M (2024) Cold fusion: Calibrated and ordinal latent distribution fusion for uncertainty-aware multimodal emotion recognition. IEEE Trans Pattern Anal Mach Intell 46(2):805\u2013822","DOI":"10.1109\/TPAMI.2023.3325770"},{"key":"447_CR47","doi-asserted-by":"crossref","unstructured":"Thulasidasan S, Chennupati G, Bilmes J, Bhattacharya T, Michalak S (2019) On mixup training: Improved calibration and predictive uncertainty for deep neural networks. In: Proceedings of NeurIPS","DOI":"10.2172\/1525811"},{"key":"447_CR48","doi-asserted-by":"publisher","unstructured":"Tsai YH, Bai S, Liang PP, Kolter JZ, Morency L, Salakhutdinov R (2019) Mult: Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of ACL. https:\/\/doi.org\/10.18653\/v1\/P19-1656. https:\/\/aclanthology.org\/P19-1656","DOI":"10.18653\/v1\/P19-1656"},{"key":"447_CR49","doi-asserted-by":"crossref","unstructured":"Wang, H., Liu, B., Li, C., Yang, Y., Li, T.: Learning with noisy labels for sentence-level sentiment classification. arXiv:1909.00124 (2019)","DOI":"10.18653\/v1\/D19-1655"},{"key":"447_CR50","doi-asserted-by":"publisher","first-page":"207","DOI":"10.1016\/j.knosys.2016.08.012","volume":"111","author":"Y Wang","year":"2016","unstructured":"Wang Y, Rao Y, Zhan X, Chen H, Luo M, Yin J (2016) Sentiment and emotion classification over noisy labels. Knowl-Based Syst 111:207\u2013216","journal-title":"Knowl-Based Syst"},{"key":"447_CR51","doi-asserted-by":"publisher","first-page":"105637","DOI":"10.1016\/j.engappai.2022.105637","volume":"117","author":"R Wang","year":"2023","unstructured":"Wang R, Liu X, Hao F, Chen X, Li X, Wang C, Wu Y (2023) Ada-ccfnet: adaptive weighted confidence-calibration fusion for membranous nephropathy. Eng Appl Artif Intell 117:105637. https:\/\/doi.org\/10.1016\/j.engappai.2022.105637","journal-title":"Eng Appl Artif Intell"},{"key":"447_CR52","doi-asserted-by":"publisher","first-page":"127181","DOI":"10.1016\/j.neucom.2023.127181","volume":"572","author":"Y Wang","year":"2024","unstructured":"Wang Y, He J, Wang D, Luo X et al (2024) Multimodal transformer with adaptive modality weighting for robust sentiment analysis. Neurocomputing 572:127181","journal-title":"Neurocomputing"},{"issue":"3","key":"447_CR53","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1007\/s10489-024-06113-6","volume":"55","author":"S Wang","year":"2025","unstructured":"Wang S, Ratnavelu K, Bin Shibghatullah AS (2025) Uefn: efficient uncertainty estimation fusion network for reliable multimodal sentiment analysis. Appl Intell 55(3):171","journal-title":"Appl Intell"},{"key":"447_CR54","doi-asserted-by":"publisher","first-page":"110262","DOI":"10.1016\/j.engappai.2025.110262","volume":"146","author":"H Wang","year":"2025","unstructured":"Wang H, Du Q, Xiang Y (2025) Image-text sentiment analysis based on hierarchical interaction fusion and contrast learning enhanced. Eng Appl Artif Intell 146:110262","journal-title":"Eng Appl Artif Intell"},{"issue":"10","key":"447_CR55","doi-asserted-by":"publisher","first-page":"12113","DOI":"10.1109\/TPAMI.2023.3275156","volume":"45","author":"P Xu","year":"2023","unstructured":"Xu P, Zhu X, Clifton DA (2023) Multimodal learning with transformers: a survey. IEEE Trans Pattern Anal Mach Intell 45(10):12113\u201312132. https:\/\/doi.org\/10.1109\/TPAMI.2023.3275156","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"447_CR56","doi-asserted-by":"publisher","first-page":"110021","DOI":"10.1016\/j.knosys.2022.110021","volume":"258","author":"J Ye","year":"2022","unstructured":"Ye J, Zhou J, Tian J, Wang R, Zhou J, Gui T, Huang X (2022) Sentiment-aware multimodal pre-training for multimodal sentiment analysis. Knowl-Based Syst 258:110021. https:\/\/doi.org\/10.1016\/j.knosys.2022.110021","journal-title":"Knowl-Based Syst"},{"key":"447_CR57","doi-asserted-by":"publisher","unstructured":"Yu J, Jiang J, et al. (2019) Adapting bert for target-oriented multimodal sentiment classification. In: Proceedings of the 28th International Joint Conference on Artificial Intelligence (IJCAI). pp 5408\u20135414. https:\/\/doi.org\/10.24963\/ijcai.2019\/751","DOI":"10.24963\/ijcai.2019\/751"},{"key":"447_CR58","doi-asserted-by":"publisher","unstructured":"Yu F, Zhang J, Zhang S, et al. (2021) Learning modality-specific representations with self-supervised objectives for multimodal sentiment (self-mm). In: Proceedings of ACM Multimedia. https:\/\/doi.org\/10.1145\/3474085.3475642","DOI":"10.1145\/3474085.3475642"},{"issue":"3","key":"447_CR59","doi-asserted-by":"publisher","first-page":"1359","DOI":"10.1007\/s00371-023-02854-6","volume":"40","author":"J Zhang","year":"2024","unstructured":"Zhang J, Qin Q, Liu X, Ye Q, Du W (2024) Emotion-wise feature interaction analysis-based visual emotion distribution learning. Vis Comput 40(3):1359\u20131368","journal-title":"Vis Comput"},{"key":"447_CR60","doi-asserted-by":"crossref","unstructured":"Zhang S, Liu J, Jiao Y, Zhang Y, Chen L, Li K (2025) A multimodal semantic fusion network with cross-modal alignment for multimodal sentiment analysis. ACM Trans Multim Comput Commun Appl","DOI":"10.1145\/3744648"},{"key":"447_CR61","doi-asserted-by":"publisher","unstructured":"Zhao F, Zhang C, Geng B (2024) Deep multimodal data fusion. ACM Comput Survey 56(9). https:\/\/doi.org\/10.1145\/3649447","DOI":"10.1145\/3649447"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-025-00447-y","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00447-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00447-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T11:26:39Z","timestamp":1773141999000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-025-00447-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,24]]},"references-count":61,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["447"],"URL":"https:\/\/doi.org\/10.1007\/s44443-025-00447-y","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,24]]},"assertion":[{"value":"4 September 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest"}}],"article-number":"89"}}