{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T05:42:36Z","timestamp":1761370956426,"version":"build-2065373602"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,8,29]],"date-time":"2025-08-29T00:00:00Z","timestamp":1756425600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,29]],"date-time":"2025-08-29T00:00:00Z","timestamp":1756425600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62276073, 61966004","62276073, 61966004"],"award-info":[{"award-number":["62276073, 61966004","62276073, 61966004"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Innovation Project of Guangxi Graduate Education","award":["No. YCBZ2024115","No. YCBZ2024115"],"award-info":[{"award-number":["No. YCBZ2024115","No. YCBZ2024115"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cogn Comput"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s12559-025-10499-x","type":"journal-article","created":{"date-parts":[[2025,8,29]],"date-time":"2025-08-29T14:01:54Z","timestamp":1756476114000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Knowledge Enhanced and Incongruity Perceiving Network for Multimodal Sarcasm Detection"],"prefix":"10.1007","volume":"17","author":[{"given":"Mingqi","family":"Liu","sequence":"first","affiliation":[]},{"given":"Zhixin","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,29]]},"reference":[{"key":"10499_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.128874","volume":"616","author":"L Ou","year":"2024","unstructured":"Ou L, Li Z. Modeling inter-modal incongruous sentiment expressions for multi-modal sarcasm detection. Neurocomputing. 2024;616:128874.","journal-title":"Neurocomputing."},{"key":"10499_CR2","doi-asserted-by":"crossref","unstructured":"Yu Z, Jin D, Wang X, Li Y, Wang L, Dang J. Commonsense knowledge enhanced sentiment dependency graph for sarcasm detection. In: Proceedings of the 32nd International Joint Conference on Artificial Intelligence. 2023. pp. 2423\u20132431.","DOI":"10.24963\/ijcai.2023\/269"},{"key":"10499_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2023.110125","volume":"136","author":"J Hua","year":"2023","unstructured":"Hua J, Cui X, Li X, Tang K, Zhu P. Multimodal fake news detection through data augmentation-based contrastive learning. Applied Soft Computing. 2023;136:110125.","journal-title":"Applied Soft Computing."},{"key":"10499_CR4","doi-asserted-by":"crossref","unstructured":"Wen CS, Jia G, Yang J. Dip: Dual incongruity perceiving network for sarcasm detection. In: Proceedings of the 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2023. pp. 2540\u20132550.","DOI":"10.1109\/CVPR52729.2023.00250"},{"key":"10499_CR5","doi-asserted-by":"crossref","unstructured":"Tang B, Lin B, Yan H, Li S. Leveraging generative large language models with visual instruction and demonstration retrieval for multimodal sarcasm detection. In: Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics. 2024. pp. 1732\u20131742.","DOI":"10.18653\/v1\/2024.naacl-long.97"},{"key":"10499_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11704-022-2256-5","volume":"17","author":"Y Zeng","year":"2023","unstructured":"Zeng Y, Li Z, Chen Z, Ma H. Aspect-level sentiment analysis based on semantic heterogeneous graph convolutional network. Frontiers of Computer Science. 2023;17:1\u201313.","journal-title":"Frontiers of Computer Science."},{"key":"10499_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.119240","volume":"213","author":"Y Zeng","year":"2022","unstructured":"Zeng Y, Li Z, Tang Z, Chen Z, Ma H. Heterogeneous graph convolution based on in-domain self-supervision for multimodal sentiment analysis. Expert Syst Appl. 2022;213:119240.","journal-title":"Expert Syst. Appl."},{"key":"10499_CR8","doi-asserted-by":"crossref","unstructured":"Yu W, Xu H, Meng F, Zhu Y, Ma Y, Wu J, Zou J, Yang K. Ch-sims: A chinese multimodal sentiment analysis dataset with fine-grained annotation of modality. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2020. pp. 3718\u20133727.","DOI":"10.18653\/v1\/2020.acl-main.343"},{"key":"10499_CR9","doi-asserted-by":"crossref","unstructured":"Yang J, Yu Niu D, Guo W, Xu Y. Confede: Contrastive feature decomposition for multimodal sentiment analysis. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics. 2023. pp. 7617\u20137630.","DOI":"10.18653\/v1\/2023.acl-long.421"},{"key":"10499_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.107335","volume":"127","author":"Y Zeng","year":"2024","unstructured":"Zeng Y, Li Z, Chen Z, Ma H. A feature-based restoration dynamic interaction network for multimodal sentiment analysis. Eng Appl Artif Intell. 2024;127:107335.","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10499_CR11","doi-asserted-by":"crossref","unstructured":"Tay Y, Luu AT, Hui SC, Su J. Reasoning with sarcasm by reading in-between. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. 2018. pp. 1010\u20131020.","DOI":"10.18653\/v1\/P18-1093"},{"key":"10499_CR12","doi-asserted-by":"crossref","unstructured":"Cai Y, Cai H, Wan X. Multi-modal sarcasm detection in twitter with hierarchical fusion model. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019. pp. 2506\u20132515.","DOI":"10.18653\/v1\/P19-1239"},{"key":"10499_CR13","doi-asserted-by":"crossref","unstructured":"Qin L, Huang S, Chen Q, Cai C, Zhang Y, Liang B, Che W, Xu R. Mmsd2.0: Towards a reliable multi-modal sarcasm detection system. In: Findings of the Association for Computational Linguistics: ACL 2023. 2023. pp. 10834\u201310845.","DOI":"10.18653\/v1\/2023.findings-acl.689"},{"key":"10499_CR14","doi-asserted-by":"crossref","unstructured":"Tian Y, Xu N, Zhang R, Mao W. Dynamic routing transformer network for multimodal sarcasm detection. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics. 2023. pp. 2468\u20132480.","DOI":"10.18653\/v1\/2023.acl-long.139"},{"key":"10499_CR15","doi-asserted-by":"crossref","unstructured":"Xie Y, Zhu Z, Chen X, Chen Z, Huang Z. Moba: Mixture of bi-directional adapter for multi-modal sarcasm detection. In: Proceedings of the 32nd ACM International Conference on Multimedia. 2024.","DOI":"10.1145\/3664647.3680914"},{"key":"10499_CR16","doi-asserted-by":"crossref","unstructured":"Tiwari D, Kanojia D, Ray A, Nunna A, Bhattacharyya P. Predict and use: Harnessing predicted gaze to improve multimodal sarcasm detection. In: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023. pp. 15933\u201315948.","DOI":"10.18653\/v1\/2023.emnlp-main.988"},{"key":"10499_CR17","unstructured":"Zhu Z, Zhuang X, Zhang Y, Xu D, Hu G, Wu X, Zheng Y. Tfcd: Towards multi-modal sarcasm detection via training-free counterfactual debiasing. Proceedings of the 33rd International Joint Conference on Artificial Intelligence. 2024. 6687\u20136695."},{"key":"10499_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102353","volume":"108","author":"H-Y Liu","year":"2024","unstructured":"Liu H-Y, Wei R, Tu G, Lin J, Liu C, Jiang D. Sarcasm driven by sentiment: A sentiment-aware hierarchical fusion network for multimodal sarcasm detection. Inf Fusion. 2024;108:102353.","journal-title":"Inf. Fusion."},{"key":"10499_CR19","doi-asserted-by":"crossref","unstructured":"Wang T, Li J, Su G, Zhang Y, Su D, Hu Y, Sha Y. Rclmufn: Relational context learning and multiplex fusion network for multimodal sarcasm detection. arXiv:2412.13008. 2024.","DOI":"10.2139\/ssrn.5108778"},{"key":"10499_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101921","volume":"100","author":"T Yue","year":"2023","unstructured":"Yue T, Mao R, Wang H, Hu Z, Cambria E. Knowlenet: Knowledge fusion network for multimodal sarcasm detection. Inf Fusion. 2023;100:101921.","journal-title":"Inf. Fusion."},{"key":"10499_CR21","doi-asserted-by":"crossref","unstructured":"Li Y, Wang Y-Z, Cui Z. Decoupled multimodal distilling for emotion recognition. In: Proceedings of the 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2023. pp. 6631\u20136640.","DOI":"10.1109\/CVPR52729.2023.00641"},{"key":"10499_CR22","doi-asserted-by":"crossref","unstructured":"Zhang C, Peng Y. Better and faster: Knowledge transfer from multiple self-supervised learning tasks via graph distillation for video classification. In: Proceedings of the 27th International Joint Conference on Artificial Intelligence. 2018","DOI":"10.24963\/ijcai.2018\/158"},{"key":"10499_CR23","doi-asserted-by":"crossref","unstructured":"Shen C, Xue M, Wang X, Song J, Sun L, Song M. Customizing student networks from heterogeneous teachers via adaptive knowledge amalgamation. In: Proceedings of the 2019 IEEE\/CVF International Conference on Computer Vision. 2019. pp. 3503\u20133512.","DOI":"10.1109\/ICCV.2019.00360"},{"key":"10499_CR24","doi-asserted-by":"crossref","unstructured":"Pontiki M, Galanis D, Papageorgiou H, Manandhar S, Androutsopoulos I, Bel N, Eryi\u01e7it G. Semeval-2016 task 5: Aspect based sentiment analysis. In: Proceedings of the 10th International Workshop on Semantic Evaluation. 2016. pp. 19\u201330.","DOI":"10.18653\/v1\/S16-1002"},{"key":"10499_CR25","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.aiopen.2023.01.001","volume":"4","author":"R Misra","year":"2023","unstructured":"Misra R, Arora P. Sarcasm detection using news headlines dataset. AI Open. 2023;4:13\u20138.","journal-title":"AI Open."},{"key":"10499_CR26","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. Imagenet: A large-scale hierarchical image database. In: Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition. 2009. pp. 248\u2013255.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"10499_CR27","doi-asserted-by":"publisher","first-page":"1452","DOI":"10.1109\/TPAMI.2017.2723009","volume":"40","author":"B Zhou","year":"2018","unstructured":"Zhou B, Lapedriza \u00c0, Khosla A, Oliva A, Torralba A. Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2018;40:1452\u201364.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence."},{"key":"10499_CR28","doi-asserted-by":"crossref","unstructured":"Achlioptas P, Ovsjanikov M, Haydarov K, Elhoseiny M, Guibas LJ. Artemis: Affective language for visual art. In: Proceedings of the 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2021. pp. 11564\u201311574.","DOI":"10.1109\/CVPR46437.2021.01140"},{"key":"10499_CR29","unstructured":"Cambria E, Liu Q, Decherchi S, Xing F, Kwok K. Senticnet 7: A commonsense-based neurosymbolic ai framework for explainable sentiment analysis. In: Proceedings of the 13th Language Resources and Evaluation Conference. 2022. pp. 3829\u20133839."},{"key":"10499_CR30","doi-asserted-by":"crossref","unstructured":"Qiao, Y., Jing, L., Song, X., Chen, X., Zhu, L., Nie, L.: Mutual-enhanced incongruity learning network for multi-modal sarcasm detection. In: Proceedings of the AAAI Conference on Artificial Intelligence. 2023. pp. 9507\u20139515.","DOI":"10.1609\/aaai.v37i8.26138"},{"key":"10499_CR31","doi-asserted-by":"crossref","unstructured":"Liu H, Wang W, Li H. Towards multi-modal sarcasm detection via hierarchical congruity modeling with knowledge enhancement. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. 2022. pp. 4995\u20135006.","DOI":"10.18653\/v1\/2022.emnlp-main.333"},{"key":"10499_CR32","doi-asserted-by":"crossref","unstructured":"Feng T, Liu J, Yang J. Probing sentiment-oriented pretraining inspired by human sentiment perception mechanism. In: Proceedings of the 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2023. pp. 2850\u20132860.","DOI":"10.1109\/CVPR52729.2023.00279"},{"key":"10499_CR33","doi-asserted-by":"publisher","first-page":"106710","DOI":"10.1016\/j.neunet.2024.106710","volume":"180","author":"J Hu","year":"2024","unstructured":"Hu J, Li Z, Su Q, Tang Z, Ma H. Exploring refined dual visual features cross-combination for image captioning. Neural networks. 2024;180:106710.","journal-title":"Neural networks"},{"key":"10499_CR34","unstructured":"Chen X, Fang H, Lin T-Y, Vedantam R, Gupta S, Doll\u00e1r P, Zitnick CL. Microsoft coco captions: Data collection and evaluation server. arXiv preprint arXiv:1504.00325. 2015."},{"key":"10499_CR35","doi-asserted-by":"crossref","unstructured":"Fu J, Xu S, Liu H, Liu Y, Xie N, Wang C, Liu J, Sun Y, Wang B. Cma-clip: Cross-modality attention clip for text-image classification. In: Proceedings of the 2022 IEEE International Conference on Image Processing. 2022. pp. 2846\u20132850.","DOI":"10.1109\/ICIP46576.2022.9897323"},{"key":"10499_CR36","doi-asserted-by":"crossref","unstructured":"Tsai Y-HH, Bai S, Liang PP, Kolter JZ, Morency L-P, Salakhutdinov R. Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019. pp. 6558\u20136569.","DOI":"10.18653\/v1\/P19-1656"},{"key":"10499_CR37","doi-asserted-by":"crossref","unstructured":"Liang B, Lou C, Li X, Yang M, Gui L, He Y, Pei W, Xu R. Multi-modal sarcasm detection via cross-modal graph convolutional network. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics. 2022. pp. 1767\u20131777.","DOI":"10.18653\/v1\/2022.acl-long.124"},{"key":"10499_CR38","doi-asserted-by":"crossref","unstructured":"Long X, Gan C, Melo G, Liu X, Li Y, Li F, Wen S. Multimodal keyless attention fusion for video classification. In: Proceedings of the AAAI Conference on Artificial Intelligence. 2018.","DOI":"10.1609\/aaai.v32i1.12319"},{"key":"10499_CR39","doi-asserted-by":"publisher","first-page":"602","DOI":"10.1016\/j.neunet.2005.06.042","volume":"18","author":"A Graves","year":"2005","unstructured":"Graves A, Schmidhuber J. Framewise phoneme classification with bidirectional lstm and other neural network architectures. Neural networks. 2005;18:602\u201310.","journal-title":"Neural networks"},{"key":"10499_CR40","doi-asserted-by":"crossref","unstructured":"Kim Y. Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014. pp. 1746\u20131751.","DOI":"10.3115\/v1\/D14-1181"},{"key":"10499_CR41","doi-asserted-by":"publisher","first-page":"2115","DOI":"10.1145\/3308558.3313735","volume":"10","author":"T Xiong","year":"2019","unstructured":"Xiong T, Zhang P, Zhu H, Yang Y. Sarcasm detection with self-matching networks and low-rank bilinear pooling. The World Wide Web Conference. 2019;10:2115\u201324.","journal-title":"The World Wide Web Conference."},{"key":"10499_CR42","unstructured":"Devlin J, Chang M-W, Lee K, Toutanova K. Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics. 2019. pp. 4171\u20134186."},{"key":"10499_CR43","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu C-Y, Feichtenhofer C, Darrell T, Xie S. A convnet for the 2020s. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2022. pp. 11966\u201311976.","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"10499_CR44","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N. An image is worth 16x16 words: Transformers for image recognition at scale. In: Proceedings of the International Conference on Learning Representations. 2021."},{"key":"10499_CR45","unstructured":"He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. arXiv preprint arxiv:1512.03385. 2015."},{"key":"10499_CR46","doi-asserted-by":"crossref","unstructured":"Xu N, Zeng Z, Mao W. Reasoning with multimodal sarcastic tweets via modeling cross-modality contrast and semantic association. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2020. pp. 3777\u20133786.","DOI":"10.18653\/v1\/2020.acl-main.349"},{"key":"10499_CR47","doi-asserted-by":"crossref","unstructured":"Pan H, Lin Z, Fu P, Qi Y, Wang W. Modeling intra and inter-modality incongruity for multi-modal sarcasm detection. In: Findings of the Association for Computational Linguistics: EMNLP 2020. 2020. pp. 1383\u20131392.","DOI":"10.18653\/v1\/2020.findings-emnlp.124"},{"key":"10499_CR48","doi-asserted-by":"crossref","unstructured":"Liang B, Lou C, Li X, Gui L, Yang M, Xu R. Multi-modal sarcasm detection with interactive in-modal and cross-modal graphs. In: Proceedings of the 29th ACM International Conference on Multimedia. 2021. pp. 4707\u20134715.","DOI":"10.1145\/3474085.3475190"},{"key":"10499_CR49","unstructured":"Ma H, He D, Wang X, Jin D, Ge M, Wang L. Multi-modal sarcasm detection based on dual generative processes. In: International Joint Conference on Artificial Intelligence. 2024."},{"key":"10499_CR50","doi-asserted-by":"crossref","unstructured":"Lin H, Chen Z, Luo Z, Cheng M, Ma J, Chen G. Cofipara: A coarse-to-fine paradigm for multimodal sarcasm target identification with large multimodal models. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics. 2024. pp. 9663\u20139687.","DOI":"10.18653\/v1\/2024.acl-long.522"},{"key":"10499_CR51","doi-asserted-by":"crossref","unstructured":"Liu H, Yang B, Yu Z. A multi-view interactive approach for multimodal sarcasm detection in social internet of things with knowledge enhancement. Applied Sciences. 2024;14(5):2146.","DOI":"10.3390\/app14052146"},{"key":"10499_CR52","doi-asserted-by":"crossref","unstructured":"Liu Y, Song R, Shi L, Gao L, Xu H. Dgja: Dependency graph-enhanced joint attention structure for multimodal sarcasm detection. In: Proceedings of the 2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2025. pp. 1\u20135.","DOI":"10.1109\/ICASSP49660.2025.10889440"},{"key":"10499_CR53","unstructured":"Chen J, Yu H, Huang S, Liu S, Zhang L. Interclip-mep: Interactive clip and memory-enhanced predictor for multi-modal sarcasm detection. arXiv preprint arxiv:2406.16464. 2024."},{"key":"10499_CR54","doi-asserted-by":"crossref","unstructured":"Li L, Jin D, Wang X, Guo F, Wang L, Dang J. Multi-modal sarcasm detection based on cross-modal composition of inscribed entity relations. In: 2023 IEEE 35th International Conference on Tools with Artificial Intelligence. 2023. pp. 918\u2013925.","DOI":"10.1109\/ICTAI59109.2023.00138"},{"key":"10499_CR55","doi-asserted-by":"crossref","unstructured":"Song L, Zhao Z, Ma Y, Liu Y, Li J. Global-aware attention network for multi-modal sarcasm detection. In: 2023 IEEE International Conference on Systems, Man, and Cybernetics. 2023. pp. 2409\u20132414.","DOI":"10.1109\/SMC53992.2023.10394483"},{"key":"10499_CR56","doi-asserted-by":"publisher","first-page":"1874","DOI":"10.1109\/TAFFC.2024.3380375","volume":"15","author":"B Liang","year":"2024","unstructured":"Liang B, Gui L, He Y, Cambria E, Xu R. Fusion and discrimination: A multimodal graph contrastive learning framework for multimodal sarcasm detection. IEEE Transactions on Affective Computing. 2024;15:1874\u201388.","journal-title":"IEEE Transactions on Affective Computing."},{"key":"10499_CR57","unstructured":"Ramakrishnan AA, Ramakrishnan AA, Lee D. Ironic: Coherence-aware reasoning chains for multi-modal sarcasm detection. arXiv:2505.16258. 2025."},{"key":"10499_CR58","unstructured":"Zhang Y, Zou C, Wang B, Qin J. Commander-gpt: Fully unleashing the sarcasm detection capability of multi-modal large language models. arxiv:2503.18681v3. 2025."}],"container-title":["Cognitive Computation"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12559-025-10499-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s12559-025-10499-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12559-025-10499-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T05:37:31Z","timestamp":1761370651000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s12559-025-10499-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,29]]},"references-count":58,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["10499"],"URL":"https:\/\/doi.org\/10.1007\/s12559-025-10499-x","relation":{},"ISSN":["1866-9956","1866-9964"],"issn-type":[{"type":"print","value":"1866-9956"},{"type":"electronic","value":"1866-9964"}],"subject":[],"published":{"date-parts":[[2025,8,29]]},"assertion":[{"value":"19 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics Approval"}},{"value":"Not applicable","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Participate"}},{"value":"Not applicable","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for Publication"}},{"value":"The authors declare no competing interests.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}],"article-number":"140"}}