{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:36:59Z","timestamp":1772908619076,"version":"3.50.1"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T00:00:00Z","timestamp":1736985600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T00:00:00Z","timestamp":1736985600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s00530-024-01656-7","type":"journal-article","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T16:50:12Z","timestamp":1737046212000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["CLIP guided image caption decoding based on monte carlo tree search"],"prefix":"10.1007","volume":"31","author":[{"given":"GuangSheng","family":"Luo","sequence":"first","affiliation":[]},{"given":"ZhiJun","family":"Fang","sequence":"additional","affiliation":[]},{"given":"JianLing","family":"Liu","sequence":"additional","affiliation":[]},{"given":"YiFanBai","family":"Bai","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,16]]},"reference":[{"key":"1656_CR1","doi-asserted-by":"crossref","unstructured":"Agrawal, H., Desai, K., Wang, Y., Chen, X., Jain, R., John-son, M., Batra, D., Parikh, D., Lee, S., Anderson, P.: nocaps: novel object captioning at scale. In: Proceed-ings of the IEEE\/CVF International Conference on Com-puter Vision, pp. 8948\u20138957 (2019)","DOI":"10.1109\/ICCV.2019.00904"},{"key":"1656_CR2","doi-asserted-by":"crossref","unstructured":"Anderson, P., Fernando, B., Johnson, M., Gould, S.: Spice: Semantic propositional image caption evalua-tion. In: European conference on computer vision, pp. 382\u2013398. Springer (2016)","DOI":"10.1007\/978-3-319-46454-1_24"},{"issue":"1","key":"1656_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TCIAIG.2012.2186810","volume":"4","author":"CB Browne","year":"2012","unstructured":"Browne, C.B., et al.: A survey of monte carlo tree search methods. IEEE Trans. Comput. Intell. AI Games 4(1), 1\u201343 (2012)","journal-title":"IEEE Trans. Comput. Intell. AI Games"},{"key":"1656_CR4","doi-asserted-by":"crossref","unstructured":"HESHAM, Mohammad, et\u00a0al.: Smart trailer: Automatic generation of movie trailer using only subtitles. In: 2018 First International Workshop on Deep and Representation Learning (IWDRL). IEEE, pp. 26\u201330 (2018)","DOI":"10.1109\/IWDRL.2018.8358211"},{"key":"1656_CR5","doi-asserted-by":"crossref","unstructured":"Cornia, M., Stefanini, M., Baraldi, L., Cucchiara, R.: Meshed-memory transformer for image captioning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10578\u201310587 (2020)","DOI":"10.1109\/CVPR42600.2020.01059"},{"key":"1656_CR6","doi-asserted-by":"crossref","unstructured":"Fei, J., Wang, T., Zhang, J., et\u00a0al.: Transferable decoding with visual entities for zero-shot image captioning. Proceedings of the IEEE\/CVF International Conference on Computer Vision pp. 3136\u20133146 (2023)","DOI":"10.1109\/ICCV51070.2023.00291"},{"key":"1656_CR7","unstructured":"Li, W., Zhu, L., Wen, L., et\u00a0al.: Decap: Decoding clip latents for zero-shot captioning via text-only training. arXiv preprint arXiv:2303.03032 (2023)"},{"key":"1656_CR8","doi-asserted-by":"crossref","unstructured":"Das, A., Kottur, S., Gupta, K., Singh, A., Yadav, D., Moura, J. M., Parikh, D., Batra, D. Visual dialog. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 326\u2013335 (2017)","DOI":"10.1109\/CVPR.2017.121"},{"key":"1656_CR9","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et\u00a0al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"1656_CR10","doi-asserted-by":"crossref","unstructured":"Fabre, B., Urvoy, T., Chevelu, J., Lolive, D.: Neural-Driven Search-Based Paraphrase Generation. In: Proceedings of the 16th Conference of the European Chap-ter of the Association for Computational Linguistics: Main Volume, pp. 2100\u20132111 (2021)","DOI":"10.18653\/v1\/2021.eacl-main.180"},{"key":"1656_CR11","doi-asserted-by":"crossref","unstructured":"Hessel, J., Holtzman, A., Forbes, M., Bras, R. L., Choi, Y.: CLIPScore: A Reference-free Evaluation Metric for Image Captioning. arXiv preprint arXiv:2104.08718 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.595"},{"key":"1656_CR12","unstructured":"Holtzman, A., Buys, J., Forbes, M., Choi, Y.: The Curious Case of Neural Text Degeneration. arXiv:1904.09751 (2020)"},{"issue":"6","key":"1656_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3295748","volume":"51","author":"MZ Hossain","year":"2019","unstructured":"Hossain, M.Z., Sohel, F., Shiratuddin, M.F., Laga, H.: A comprehensive survey of deep learning for image captioning. ACM Comput. Surveys (CsUR) 51(6), 1\u201336 (2019)","journal-title":"ACM Comput. Surveys (CsUR)"},{"key":"1656_CR14","doi-asserted-by":"crossref","unstructured":"Karpathy, A., Fei-Fei, L.: Deep visual-semantic alignments for generating image descriptions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3128\u20133137 (2015)","DOI":"10.1109\/CVPR.2015.7298932"},{"key":"1656_CR15","unstructured":"Mokady, R., Hertz, A., Bermano, A.H.: Clipcap: Clip prefix for image captioning. arXiv preprint arXiv:2111.09734, (2021)"},{"key":"1656_CR16","doi-asserted-by":"crossref","unstructured":"Rennie, S.J., Marcheret, E., Mroueh, Y., et\u00a0al.: Self-critical sequence training for image captioning. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 7008\u20137024 (2017)","DOI":"10.1109\/CVPR.2017.131"},{"key":"1656_CR17","doi-asserted-by":"crossref","unstructured":"Kumagai, K., Kobayashi, I., Mochihashi, D., Asoh, H., Nakamura, T., Nagai, T.: Human-like natural lan-guage generation using monte carlo tree search. In: Proceed-ings of the INLG 2016 Workshop on Computational Creativ-ity in Natural Language Generation, pp. 11\u201318 (2016)","DOI":"10.18653\/v1\/W16-5502"},{"key":"1656_CR18","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ra-manan, D., Doll\u00e1r, P., and Zitnick, C. L. 2014. Microsoftcoco: Common objects in context. In: European conference on computer vision, pp. 740\u2013755, Springer","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1656_CR19","doi-asserted-by":"crossref","unstructured":"Liu, S., Zhu, Z., Ye, N., Guadarrama, S., Murphy, K.: Improved image captioning via policy gradient opti-mization of spider. In: Proceedings of the IEEE international conference on computer vision, pp. 873\u2013881, (2017)","DOI":"10.1109\/ICCV.2017.100"},{"key":"1656_CR20","doi-asserted-by":"crossref","unstructured":"Luo, Y., Ji, J., Sun, X., Cao, L., Wu, Y., Huang, F., Lin, C.-W., Ji, R.: Dual-level collaborative transformer for image captioning. arXiv preprint arXiv:2101.06462, (2021)","DOI":"10.1609\/aaai.v35i3.16328"},{"key":"1656_CR21","unstructured":"Luo, R.: A Better Variant of Self-Critical Sequence Training. arXiv preprint arXiv:2003.09971, (2020)"},{"key":"1656_CR22","doi-asserted-by":"crossref","unstructured":"Rennie, S.J., Marcheret, E., Mroueh, Y., Ross, J., Goel, V.: Self-critical sequence training for image caption-ing. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7008\u20137024 (2017)","DOI":"10.1109\/CVPR.2017.131"},{"key":"1656_CR23","unstructured":"Scialom, T., Dray, P.-A., Lamprier, S., Piwowarski, B., Staiano, J.: To Beam Or Not To Beam: That is a Ques-tion of Cooperation for Language GANs. arXiv preprint arXiv:2106.06363 (2021)"},{"issue":"7676","key":"1656_CR24","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1038\/nature24270","volume":"550","author":"D Silver","year":"2017","unstructured":"Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L., Lai, M., Bolton, A., et al.: Mastering the game of go without human knowledge. Nature 550(7676), 354\u2013359 (2017)","journal-title":"Nature"},{"key":"1656_CR25","unstructured":"Stefanini, M., Cornia, M., Baraldi, L., Cascianelli, S., Fiameni, G., Cucchiara, R.: From Show to Tell: A Survey on Image Captioning. arXiv preprint arXiv:2107.06912 (2021)"},{"key":"1656_CR26","unstructured":"Sutton, R.S., Barto, A.G.: Reinforcement learn-ing: An introduction. MIT press (2018)"},{"key":"1656_CR27","unstructured":"Vaswani, A., Shazeer, N. M., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., Polosukhin, I.: Attention is All you Need. arXiv:1706.03762 (2017)"},{"key":"1656_CR28","doi-asserted-by":"crossref","unstructured":"Vedantam, R., Lawrence Zitnick, C., Parikh, D.: Cider: Consensus-based image description evaluation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 4566\u20134575 (2015)","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"1656_CR29","doi-asserted-by":"crossref","unstructured":"Wu, S., Wieland, J., Farivar, O., Schiller, J.: Au-tomatic alt-text: Computer-generated image descriptions for blind users on a social network service. In: Proceedings of the 2017 ACM Conference on Computer Supported Cooper-ative Work and Social Computing, pp. 1180\u20131192 (2017)","DOI":"10.1145\/2998181.2998364"},{"key":"1656_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, P., Li, X., Hu, X., Yang, J., Zhang, L., Wang, L., Choi, Y., Gao, J. Vinvl: Revisiting visual rep-resentations in vision-language models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5579\u20135588 (2021)","DOI":"10.1109\/CVPR46437.2021.00553"},{"key":"1656_CR31","doi-asserted-by":"crossref","unstructured":"Zhou, L., Palangi, H., Zhang, L., Hu, H., Corso, J.J., Gao, J.: Unified Vision-Language Pre-Training for Im-age Captioning and VQA. In AAAI (2020)","DOI":"10.1609\/aaai.v34i07.7005"},{"issue":"10","key":"1656_CR32","doi-asserted-by":"publisher","first-page":"2473","DOI":"10.1109\/TCSVT.2017.2706264","volume":"28","author":"J Han","year":"2017","unstructured":"Han, J., Cheng, G., Li, Z., et al.: A unified metric learning-based framework for co-saliency detection. IEEE Trans. Circuits Syst. Video Technol. 28(10), 2473\u20132483 (2017)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"4","key":"1656_CR33","doi-asserted-by":"publisher","first-page":"1258","DOI":"10.1007\/s11263-020-01419-7","volume":"129","author":"K Ding","year":"2021","unstructured":"Ding, K., Ma, K., Wang, S., et al.: Comparison of full-reference image quality models for optimization of image processing systems. Int. J. Comput. Vision 129(4), 1258\u20131281 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"1656_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.128666","volume":"611","author":"J Yan","year":"2025","unstructured":"Yan, J., Xie, Y., Zou, S., et al.: EntroCap: Zero-shot image captioning with entropy-based retrieval. Neurocomputing 611, 128666 (2025)","journal-title":"Neurocomputing"},{"key":"1656_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2024.106813","volume":"181","author":"L Chen","year":"2025","unstructured":"Chen, L., Li, K.: Multi-Modal Graph Aggregation Transformer for image captioning. Neural Netw. 181, 106813 (2025)","journal-title":"Neural Netw."},{"key":"1656_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2024.110941","volume":"158","author":"Y Li","year":"2025","unstructured":"Li, Y., Ji, J., Sun, X., et al.: M3ixup: A multi-modal data augmentation approach for image captioning. Pattern Recognit. 158, 110941 (2025)","journal-title":"Pattern Recognit."},{"key":"1656_CR37","unstructured":"Elguendouze, S.: Explainable Artificial Intelligence approaches for Image Captioning. Universit\u00e9 d\u2019Orl\u00e9ans (2024)"},{"key":"1656_CR38","doi-asserted-by":"crossref","unstructured":"Sarto, S., Moratelli, N., Cornia, M., et\u00a0al.: Positive-Augmented Contrastive Learning for Vision-and-Language Evaluation and Training. arXiv preprint arXiv:2410.07336 (2024)","DOI":"10.1109\/CVPR52729.2023.00668"},{"key":"1656_CR39","doi-asserted-by":"crossref","unstructured":"Sarto, S., Barraco, M., Cornia, M., et\u00a0al.: Positive-augmented contrastive learning for image and video captioning evaluation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 6914\u20136924 (2023)","DOI":"10.1109\/CVPR52729.2023.00668"},{"key":"1656_CR40","first-page":"79124","volume":"36","author":"Z Yue","year":"2023","unstructured":"Yue, Z., Hu, A., Zhang, L., et al.: Learning descriptive image captioning via semipermeable maximum likelihood estimation. Adv. Neural Inf. Process. Syst. 36, 79124\u201379141 (2023)","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"2","key":"1656_CR41","first-page":"4","volume":"2","author":"M Magnuson","year":"2015","unstructured":"Magnuson, M.: Monte carlo tree search and its applications. Scholarly Horizons: University of Minnesota. Morris Undergr. J. 2(2), 4 (2015)","journal-title":"Morris Undergr. J."},{"key":"1656_CR42","doi-asserted-by":"crossref","unstructured":"Kumagai, K, Kobayashi, I., Mochihashi, D., et\u00a0al.: Human-like natural language generation using monte carlo tree search. In: Proceedings of the INLG 2016 Workshop on Computational Creativity in Natural Language Generation pp. 11\u201318 (2016)","DOI":"10.18653\/v1\/W16-5502"},{"issue":"1","key":"1656_CR43","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TCIAIG.2012.2186810","volume":"4","author":"CB Browne","year":"2012","unstructured":"Browne, C.B., Powley, E., Whitehouse, D., et al.: A survey of monte carlo tree search methods. IEEE Trans. Comput. Intell. AI games 4(1), 1\u201343 (2012)","journal-title":"IEEE Trans. Comput. Intell. AI games"},{"key":"1656_CR44","first-page":"15908","volume":"34","author":"K Han","year":"2021","unstructured":"Han, K., Xiao, A., Wu, E., et al.: Transformer in transformer. Adv. Neural Inf. Process. Syst. 34, 15908\u201315919 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1656_CR45","unstructured":"Li, Y.: Deep reinforcement learning: An overview (2017). arXiv preprint arXiv:1701.07274"},{"issue":"12","key":"1656_CR46","first-page":"10887","volume":"35","author":"J Zhang","year":"2021","unstructured":"Zhang, J., Kim, J., O\u2019Donoghue, B., et al.: Sample efficient reinforcement learning with Reinforce. Proc. AAAI Conf. Artif. Intell. 35(12), 10887\u201310895 (2021)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"1656_CR47","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et\u00a0al.: Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020 (2021)"},{"key":"1656_CR48","doi-asserted-by":"crossref","unstructured":"Carpentier A, Lazaric A, Ghavamzadeh M, et\u00a0al. Upper-confidence-bound algorithms for active learning in multi-armed bandits. In: International Conference on Algorithmic Learning Theory. Springer, Berlin, Heidelberg pp. 189\u2013203 (2011)","DOI":"10.1007\/978-3-642-24412-4_17"},{"key":"1656_CR49","doi-asserted-by":"crossref","unstructured":"Matsuzaki K. Empirical analysis of PUCT algorithm with evaluation functions of different quality. In: 2018 Conference on Technologies and Applications of Artificial Intelligence (TAAI). IEEE, pp. 142\u2013147 (2018)","DOI":"10.1109\/TAAI.2018.00043"},{"issue":"4","key":"1656_CR50","first-page":"36","volume":"9","author":"JS Lee","year":"2020","unstructured":"Lee, J.S., Oh, I.S.: Hyper-parameter Optimization for Monte Carlo Tree Search using Self-play. Smart Media J. 9(4), 36\u201343 (2020)","journal-title":"Smart Media J."},{"key":"1656_CR51","doi-asserted-by":"crossref","unstructured":"Shaham U, Levy O. What Do You Get When You Cross Beam Search with Nucleus Sampling?. arXiv preprint arXiv:2107.09729, (2021)","DOI":"10.18653\/v1\/2022.insights-1.5"},{"key":"1656_CR52","doi-asserted-by":"publisher","DOI":"10.1016\/j.cie.2021.107211","volume":"155","author":"K Li","year":"2021","unstructured":"Li, K., Deng, Q., Zhang, L., et al.: An effective MCTS-based algorithm for minimizing makespan in dynamic flexible job shop scheduling problem. Comput. Indust. Eng. 155, 107211 (2021)","journal-title":"Comput. Indust. Eng."},{"key":"1656_CR53","unstructured":"Mokady R, Hertz A, Bermano A H. Clipcap: Clip prefix for image captioning. arXiv preprint arXiv:2111.09734 (2021)"},{"key":"1656_CR54","doi-asserted-by":"crossref","unstructured":"Kuo C W, Kira Z. Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for Image Captioning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 17969\u201317979 (2022)","DOI":"10.1109\/CVPR52688.2022.01744"},{"key":"1656_CR55","doi-asserted-by":"crossref","unstructured":"Arnab A, Dehghani M, Heigold G, et\u00a0al. Vivit: A video vision transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 6836\u20136846 (2021)","DOI":"10.1109\/ICCV48922.2021.00676"},{"key":"1656_CR56","doi-asserted-by":"crossref","unstructured":"Zhang B, Titov I, Sennrich R. Improving deep transformer with depth-scaled initialization and merged attention. arXiv preprint arXiv:1908.11365 (2019)","DOI":"10.18653\/v1\/D19-1083"},{"key":"1656_CR57","doi-asserted-by":"crossref","unstructured":"Barraco M, Cornia M, Cascianelli S, et\u00a0al. The Unreasonable Effectiveness of CLIP Features for Image Captioning: An Experimental Analysis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 4662\u20134670 (2022)","DOI":"10.1109\/CVPRW56347.2022.00512"},{"key":"1656_CR58","unstructured":"Veit A, Matera T, Neumann L, et\u00a0al. Coco-text: Dataset and benchmark for text detection and recognition in natural images. arXiv preprint arXiv:1601.07140 (2016)"},{"key":"1656_CR59","doi-asserted-by":"crossref","unstructured":"Rennie S J, Marcheret E, Mroueh Y, et\u00a0al. Self-critical sequence training for image captioning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7008\u20137024 (2017)","DOI":"10.1109\/CVPR.2017.131"},{"key":"1656_CR60","doi-asserted-by":"crossref","unstructured":"Hu J C, Cavicchioli R, Capotondi A. Exploiting Multiple Sequence Lengths in Fast End to End Training for Image Captioning. In: 2023 IEEE International Conference on Big Data (BigData). IEEE pp. 2173\u20132182 (2023)","DOI":"10.1109\/BigData59044.2023.10386812"},{"key":"1656_CR61","unstructured":"Wang Z, Yu J, Yu A W, et\u00a0al. Simvlm: Simple visual language model pretraining with weak supervision (2021). arXiv preprint arXiv:2108.10904"},{"key":"1656_CR62","unstructured":"Yu J, Wang Z, Vasudevan V, et\u00a0al. Coca: Contrastive captioners are image-text foundation models, (2022). arXiv preprint arXiv:2205.01917"},{"key":"1656_CR63","unstructured":"Chang W C, Yu H F, Zhong K, et\u00a0al. A modular deep learning approach for extreme multi-label text classification. (2019)"},{"key":"1656_CR64","doi-asserted-by":"crossref","unstructured":"Zhou L, Palangi H, Zhang L, et\u00a0al. Unified vision-language pre-training for image captioning and vqa. In: Proceedings of the AAAI Conference on Artificial Intelligence 34(07): 13041\u201313049 (2020)","DOI":"10.1609\/aaai.v34i07.7005"},{"key":"1656_CR65","doi-asserted-by":"crossref","unstructured":"Zhang P, Li X, Hu X, et\u00a0al. Vinvl: Revisiting visual representations in vision-language models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 5579\u20135588 (2021)","DOI":"10.1109\/CVPR46437.2021.00553"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01656-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01656-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01656-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,28]],"date-time":"2025-02-28T11:08:32Z","timestamp":1740740912000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01656-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,16]]},"references-count":65,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["1656"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01656-7","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,16]]},"assertion":[{"value":"25 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"72"}}