{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T18:38:30Z","timestamp":1772044710644,"version":"3.50.1"},"reference-count":69,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T00:00:00Z","timestamp":1748044800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T00:00:00Z","timestamp":1748044800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100014940","name":"Chinese Polar Environment Comprehensive Investigation and Assessment Programmes","doi-asserted-by":"publisher","award":["2022YFB4500800"],"award-info":[{"award-number":["2022YFB4500800"]}],"id":[{"id":"10.13039\/501100014940","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s11263-025-02451-1","type":"journal-article","created":{"date-parts":[[2025,5,24]],"date-time":"2025-05-24T04:33:43Z","timestamp":1748061223000},"page":"5822-5839","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Supplementary Prompt Learning for Vision-Language Models"],"prefix":"10.1007","volume":"133","author":[{"given":"Rongfei","family":"Zeng","sequence":"first","affiliation":[]},{"given":"Zhipeng","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Ruiyun","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Yonggang","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,24]]},"reference":[{"key":"2451_CR1","unstructured":"Bahng, H., Jahanian, A., Sankaranarayanan, S., et\u00a0al. (2022). Exploring visual prompts for adapting large-scale models. arXiv:2203.17274"},{"key":"2451_CR2","first-page":"25005","volume":"35","author":"A Bar","year":"2022","unstructured":"Bar, A., Gandelsman, Y., Darrell, T., et al. (2022). Visual prompting via image inpainting. Advances in Neural Information Processing Systems (NeurIPS), 35, 25005\u201325017.","journal-title":"Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"2451_CR3","doi-asserted-by":"crossref","unstructured":"Berg, T., Liu, J., Woo\u00a0Lee, S., et\u00a0al. (2014). Birdsnap: Large-scale fine-grained visual categorization of birds. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 2011\u20132018).","DOI":"10.1109\/CVPR.2014.259"},{"key":"2451_CR4","doi-asserted-by":"crossref","unstructured":"Bossard, L., Guillaumin, M., & VanGool, L. (2014). Food-101\u2014Mining discriminative components with random forests. In Proceedings of the European conference on computer vision (ECCV) (pp. 446\u2013461).","DOI":"10.1007\/978-3-319-10599-4_29"},{"key":"2451_CR5","unstructured":"Bridle, J., Heading, A., & MacKay, D. (1991). Unsupervised classifiers, mutual information and \u2019phantom targets. Advances in Neural Information Processing systems 4"},{"issue":"6","key":"2451_CR6","doi-asserted-by":"publisher","first-page":"737","DOI":"10.1016\/j.ccell.2017.05.005","volume":"31","author":"F Cavalli","year":"2017","unstructured":"Cavalli, F., Remke, M., Rampasek, L., et al. (2017). Intertumoral heterogeneity within medulloblastoma subgroups. Cancer Cell, 31(6), 737\u2013754.","journal-title":"Cancer Cell"},{"key":"2451_CR7","first-page":"1327","volume":"3","author":"Y Chen","year":"2022","unstructured":"Chen, Y., Mancini, M., Zhu, X., et al. (2022). Semi-supervised and unsupervised deep visual learning: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 3, 1327\u20131347.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"},{"issue":"10","key":"2451_CR8","doi-asserted-by":"publisher","first-page":"1865","DOI":"10.1109\/JPROC.2017.2675998","volume":"105","author":"G Cheng","year":"2017","unstructured":"Cheng, G., Han, J., & Lu, X. (2017). Remote sensing image scene classification: Benchmark and state of the art. Proceedings of the IEEE, 105(10), 1865\u20131883.","journal-title":"Proceedings of the IEEE"},{"key":"2451_CR9","doi-asserted-by":"crossref","unstructured":"Cimpoi, M., Maji, S., Kokkinos, I., et\u00a0al. (2014) Describing textures in the wild. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 3606\u20133613).","DOI":"10.1109\/CVPR.2014.461"},{"key":"2451_CR10","unstructured":"Dan, H., Kevin, Z., Steven, B., et\u00a0al. (2021a). Natural adversarial examples. In Proceeding of IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 15262\u201315271)."},{"key":"2451_CR11","unstructured":"Dan, H., Steven, B., Norman, M., et\u00a0al. (2021b). The many faces of robustness: A critical analysis of out-of-distribution generalization. In Proceedings of IEEE\/CVF international conference on computer vision (ICCV) (pp. 8340\u20138349)."},{"key":"2451_CR12","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., et\u00a0al. (2009). Imagenet: A large-scale hierarchical image database. In Proceeding of IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 248\u2013255).","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"2451_CR13","doi-asserted-by":"crossref","unstructured":"Du, Y., Wei, F., Zhang, Z., et\u00a0al. (2022). Learning to prompt for open-vocabulary object detection with vision-language model. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 14084\u201314093).","DOI":"10.1109\/CVPR52688.2022.01369"},{"key":"2451_CR14","doi-asserted-by":"crossref","unstructured":"Feng, C. M., Yu, K., Liu, Y., et\u00a0al. (2023). Diverse data augmentation with diffusions for effective test-time prompt tuning. arXiv:2308.06038","DOI":"10.1109\/ICCV51070.2023.00255"},{"key":"2451_CR15","doi-asserted-by":"crossref","unstructured":"Gao, J., Zhang, J., Liu, X., et\u00a0al. (2022). Back to the source: Diffusion-driven test-time adaptation. arXiv:2207.03442","DOI":"10.1109\/CVPR52729.2023.01134"},{"key":"2451_CR16","first-page":"1","volume":"132","author":"P Gao","year":"2023","unstructured":"Gao, P., Geng, S., Zhang, R., et al. (2023). Clip-adapter: Better vision-language models with feature adapters. International Journal of Computer Vision (IJCV), 132, 1\u201315.","journal-title":"International Journal of Computer Vision (IJCV)"},{"key":"2451_CR17","doi-asserted-by":"crossref","unstructured":"Goodfellow, I., Erhan, D., Carrier, P., et\u00a0al. (2013). Challenges in representation learning: A report on three machine learning contests. In 20th international conference on neural information processing (ICONIP) (pp. 117\u2013124).","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"2451_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102136","volume":"72","author":"Y He","year":"2021","unstructured":"He, Y., Carass, A., Zuo, L., et al. (2021). Autoencoder-based self-supervised test-time adaptation for medical image analysis. Elsevier Medical image analysis, 72, Article 102136.","journal-title":"Elsevier Medical image analysis"},{"issue":"7","key":"2451_CR19","doi-asserted-by":"publisher","first-page":"2217","DOI":"10.1109\/JSTARS.2019.2918242","volume":"12","author":"P Helber","year":"2019","unstructured":"Helber, P., Bischke, B., Dengel, A., et al. (2019). Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 12(7), 2217\u20132226.","journal-title":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing"},{"key":"2451_CR20","unstructured":"Huang, T., Chu, J., & Wei, F. (2022). Unsupervised prompt learning for vision-language models. arXiv:2204.03649"},{"key":"2451_CR21","unstructured":"Jang, M., Chung, S. Y., Chung, H. W. (2022). Test-time adaptation via self-training with nearest neighbor information. arXiv:2207.10792"},{"key":"2451_CR22","unstructured":"Jia, C., Yang, Y., Xia, Y., et\u00a0al. (2021). Align: Scaling up visual and vision-language representation learning with noisy text supervision. arXiv:2102.05918"},{"key":"2451_CR23","doi-asserted-by":"crossref","unstructured":"Jia, M., Tang, L., Chen, B. C., et\u00a0al. (2022). Visual prompt tuning. In European conference on computer vision (ECCV) (pp. 709\u2013727).","DOI":"10.1007\/978-3-031-19827-4_41"},{"key":"2451_CR24","unstructured":"Jian, L., Hu, D., & Jiashi, F. (2020). Do we really need to access the source data. In International conference on machine learning (ICML) (pp. 6028\u20136039)."},{"key":"2451_CR25","unstructured":"Jonathan, K., Michael, S., Jia, D., et\u00a0al. (2013). 3D object representations for fine-grained categorization. In Proceedings of the IEEE 4th international workshop on 3D representation and recognition (pp. 554\u2013561)."},{"key":"2451_CR26","doi-asserted-by":"crossref","unstructured":"Ju, C., Han, T., Zheng, K., et\u00a0al. (2022). Prompting visual-language models for efficient video understanding. In European conference on computer vision (ECCV) (pp. 105\u2013124). Springer.","DOI":"10.1007\/978-3-031-19833-5_7"},{"key":"2451_CR27","unstructured":"Junnan, L., Dongxu, L., Silvio, S., et\u00a0al. (2023). BLIP-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In PMLR international conference on machine learning (ICML) (pp. 1\u20138)."},{"key":"2451_CR28","doi-asserted-by":"crossref","unstructured":"Khattak, M., Rasheed, H., Maaz, M., et\u00a0al. (2023). Maple: Multi-modal prompt learning. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 19113\u201319122).","DOI":"10.1109\/CVPR52729.2023.01832"},{"key":"2451_CR29","unstructured":"Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images. Technical Report."},{"key":"2451_CR30","unstructured":"Li, F. F., Fergus, R., & Perona, P. (2004). Learning generative visual models from few training xxamples: An incremental Bayesian approach tested on 101 object categories. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition workshop (CVPRW) (pp. 178\u2013178)."},{"key":"2451_CR31","unstructured":"Li, J., Li, D., Xiong, C., et\u00a0al. (2022). Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In PMLR international conference on machine learning (ICML) (pp. 12888\u201312900)."},{"key":"2451_CR32","doi-asserted-by":"crossref","unstructured":"Liang, J., Hu, D., Feng, J., et\u00a0al. (2022). Dine: Domain adaptation from single and multiple black-box predictors. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 8003\u20138013).","DOI":"10.1109\/CVPR52688.2022.00784"},{"key":"2451_CR33","unstructured":"Lim, H., Kim, B., Choo, J., et\u00a0al. (2023). Ttn: A domain-shift aware batch normalization in test-time adaptation. arXiv:2302.05155"},{"key":"2451_CR34","unstructured":"Lin, J., & Gong, S. (2023). Gridclip: One-stage object detection by grid-level clip representation learning. arXiv:2303.09252"},{"issue":"3","key":"2451_CR35","doi-asserted-by":"publisher","first-page":"645","DOI":"10.1109\/TMM.2017.2751966","volume":"20","author":"X Liu","year":"2017","unstructured":"Liu, X., Liu, W., Mei, T., et al. (2017). Provid: Progressive and multimodal vehicle reidentification for large-scale urban surveillance. IEEE Transactions on Multimedia, 20(3), 645\u2013658.","journal-title":"IEEE Transactions on Multimedia"},{"key":"2451_CR36","first-page":"21808","volume":"34","author":"Y Liu","year":"2021","unstructured":"Liu, Y., Kothari, P., Van Delft, B., et al. (2021). Ttt++: When does self-supervised test-time training fail or thrive? Advances in Neural Information Processing Systems, 34, 21808\u201321820.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2451_CR37","doi-asserted-by":"crossref","unstructured":"Lu, Y., Liu, J., Zhang, Y., et\u00a0al. (2022). Prompt distribution learning. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 5206\u20135215).","DOI":"10.1109\/CVPR52688.2022.00514"},{"key":"2451_CR38","doi-asserted-by":"crossref","unstructured":"Luddecke, T., & Ecker, A. (2022). Image segmentation using text and image prompts. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 7086\u20137096).","DOI":"10.1109\/CVPR52688.2022.00695"},{"key":"2451_CR39","unstructured":"Maji, S., Rahtu, E., Kannala, J., et\u00a0al. (2013). Fine-grained visual classification of aircraft. arXiv:1306.5151"},{"key":"2451_CR40","doi-asserted-by":"crossref","unstructured":"Nilsback, M. E., & Zisserman, A. (2006). A visual vocabulary for flower classification. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 1447\u20131454).","DOI":"10.1109\/CVPR.2006.42"},{"key":"2451_CR41","doi-asserted-by":"crossref","unstructured":"Omkar, P., Andrea, V., Andrew, Z., et\u00a0al. (2012). Cats and dogs. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 3498\u20133505).","DOI":"10.1109\/CVPR.2012.6248092"},{"key":"2451_CR42","doi-asserted-by":"crossref","unstructured":"Pandey, P., Raman, M., Varambally, S., et\u00a0al. (2021). Generalization on unseen domains via inference-time label-preserving target projections. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 12924\u201312933).","DOI":"10.1109\/CVPR46437.2021.01273"},{"key":"2451_CR43","doi-asserted-by":"crossref","unstructured":"Peng, Q., Ding, Z., Lyu, L., et\u00a0al.(2023), Rain: Regularization on input and network for black-box domain adaptation. In Proceedings of the thirty-second international joint conference on artificial intelligence (IJCAI) (pp. 4118\u20134126).","DOI":"10.24963\/ijcai.2023\/458"},{"key":"2451_CR44","unstructured":"Radford, A., Kim, J. W., Hallacy, C., et\u00a0al. (2021). Learning transferable visual models from natural language supervision. In PMLR international conference on machine learning (ICML) (pp. 8748\u20138763)."},{"key":"2451_CR45","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., et\u00a0al. (2022). Hierarchical text-conditional image generation with clip latents. arXiv:2204.06125"},{"key":"2451_CR46","unstructured":"Recht, B., Roelofs, R., Schmidt, L., et\u00a0al. (2019). Do imagenet classifiers generalize to imagenet? In PMLR international conference on machine learning (ICML) (pp. 5389\u20135400)."},{"key":"2451_CR47","doi-asserted-by":"crossref","unstructured":"Schroff, F., Kalenichenko, D., Philbin, J. (2015). Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) (pp. 815\u2013823).","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"2451_CR48","first-page":"14274","volume":"35","author":"M Shu","year":"2022","unstructured":"Shu, M., Nie, W., Huang, D. A., et al. (2022). Test-time prompt tuning for zero-shot generalization in vision-language models. Advances in Neural Information Processing Systems (NeurIPS), 35, 14274\u201314289.","journal-title":"Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"2451_CR49","doi-asserted-by":"crossref","unstructured":"Singha, M., Jha, A., Solanki, B., et\u00a0al. (2023). Applenet: Visual attention parameterized prompt learning for few-shot remote sensing image generalization using clip. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 3606\u20133613).","DOI":"10.1109\/CVPRW59228.2023.00196"},{"key":"2451_CR50","unstructured":"Soomro, K., Zamir, A.R., & Shah, M. (2012). Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv:1212.0402"},{"key":"2451_CR51","doi-asserted-by":"crossref","unstructured":"Tao, M., Bao, B. K., Tang, H., et\u00a0al. (2023). Galip: Generative adversarial clips for text-to-image synthesis. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 14214\u201314223).","DOI":"10.1109\/CVPR52729.2023.01366"},{"key":"2451_CR52","unstructured":"Wang, D., Shelhamer, E., Liu, S., et\u00a0al. (2021). Tent: Fully test-time adaptation by entropy minimization. In International conference on learning representations (ICLR) (pp. 1\u201310)."},{"key":"2451_CR53","unstructured":"Wang, H., Ge, S., Lipton, Z., et\u00a0al. (2019). Learning robust global representations by penalizing local predictive power. In Advances in neural information processing systems (NeurIPS) (pp. 10506\u201310518)."},{"key":"2451_CR54","doi-asserted-by":"crossref","unstructured":"Wu, W., Luo, H., Fang, B., et\u00a0al. (2023a). Cap4video: What can auxiliary captions do for text-video retrieval? In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 10704\u201310713).","DOI":"10.1109\/CVPR52729.2023.01031"},{"key":"2451_CR55","doi-asserted-by":"crossref","unstructured":"Wu, X., Zhu, F., Zhao, R., et\u00a0al. (2023b). Cora: Adapting clip for open-vocabulary detection with region prompting and anchor pre-matching. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 7031\u20137040).","DOI":"10.1109\/CVPR52729.2023.00679"},{"key":"2451_CR56","doi-asserted-by":"crossref","unstructured":"Wu, Z., Xiong, Y., Yu, S. X., et\u00a0al. (2018). Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3733\u20133742).","DOI":"10.1109\/CVPR.2018.00393"},{"key":"2451_CR57","doi-asserted-by":"crossref","unstructured":"Xiao, J., James, H., Krista, E., et\u00a0al. (2010). Sun database: Large-scale scene fecognition from abbey to zoo. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 3485\u20133492).","DOI":"10.1109\/CVPR.2010.5539970"},{"key":"2451_CR58","unstructured":"Yang, J., Peng, X., Wang, K., et\u00a0al. (2022). Divide to adapt: Mitigating confirmation bias for domain adaptation of black-box predictors. arXiv:2205.14467"},{"key":"2451_CR59","first-page":"9125","volume":"35","author":"L Yao","year":"2022","unstructured":"Yao, L., Han, J., Wen, Y., et al. (2022). Detclip: Dictionary-enriched visual-concept paralleled pre-training for open-world detection. Advances in Neural Information Processing Systems (NeurIPS), 35, 9125\u20139138.","journal-title":"Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"2451_CR60","unstructured":"Yuan, L., Chen, D., Chen, Y. L., et\u00a0al. (2021). Florence: A new foundation model for computer vision. arXiv:2111.11432"},{"key":"2451_CR61","unstructured":"Zhang, J., Huang, J., Jin, S., et\u00a0al. (2023). Vision-language models for vision tasks: A survey. arXiv:2304.00685"},{"key":"2451_CR62","first-page":"38629","volume":"35","author":"M Zhang","year":"2022","unstructured":"Zhang, M., Levine, S., & Finn, C. (2022). Memo: Test time robustness via adaptation and augmentation. Advances in Neural Information Processing Systems (NeurIPS), 35, 38629\u201338642.","journal-title":"Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"2451_CR63","unstructured":"Zhao, B., Chen, C., & Xia, S. T. (2023). Delta: Degradation-free fully test-time adaptation. arXiv:2301.13018"},{"key":"2451_CR64","doi-asserted-by":"crossref","unstructured":"Zhao, X., Liu, C., Sicilia, A., et\u00a0al. (2022). Test-time fourier style calibration for domain generalization. arXiv:2205.06427","DOI":"10.24963\/ijcai.2022\/240"},{"key":"2451_CR65","doi-asserted-by":"crossref","unstructured":"Zhong, Y., Yang, J., Zhang, P., et\u00a0al. (2022). Regionclip: Region-based language-image pretraining. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 16793\u201316803).","DOI":"10.1109\/CVPR52688.2022.01629"},{"key":"2451_CR66","doi-asserted-by":"crossref","unstructured":"Zhou, K., Yang, J., Loy, C. C., et\u00a0al. (2022a). Conditional prompt learning for vision-language models. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR) (pp. 16816\u201316825).","DOI":"10.1109\/CVPR52688.2022.01631"},{"key":"2451_CR67","doi-asserted-by":"publisher","first-page":"2337","DOI":"10.1007\/s11263-022-01653-1","volume":"130","author":"K Zhou","year":"2022","unstructured":"Zhou, K., Yang, J., Loy, C. C., et al. (2022). Learning to prompt for vision-language models. Proceedings of the International Journal of Computer Vision (IJCV), 130, 2337\u20132348.","journal-title":"Proceedings of the International Journal of Computer Vision (IJCV)"},{"key":"2451_CR68","unstructured":"Zhou, Y., Ram, P., Salonidis, T., et\u00a0al. (2021). Flora: Single-shot hyper-parameter optimization for federated learning. arXiv:2112.08524"},{"key":"2451_CR69","unstructured":"Zhu, D., Chen, J., Shen, X., et\u00a0al. (2023). Minigpt-4: Enhancing vision-language understanding with advanced large language models. arXiv:2304.10592"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02451-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02451-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02451-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T15:19:11Z","timestamp":1757171951000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02451-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,24]]},"references-count":69,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["2451"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02451-1","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,24]]},"assertion":[{"value":"3 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 April 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 May 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}