{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:27:10Z","timestamp":1757618830111,"version":"3.44.0"},"publisher-location":"Singapore","reference-count":27,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819697939"},{"type":"electronic","value":"9789819697946"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-9794-6_40","type":"book-chapter","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T06:12:07Z","timestamp":1752473527000},"page":"479-490","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["LoRA-CLIP: Low-Rank Adaptation of Text Prompts for Vision-Language Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-2469-1914","authenticated-orcid":false,"given":"Ping","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Haocun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Kaiyan","family":"Song","sequence":"additional","affiliation":[]},{"given":"Haizheng","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Hong","family":"Bian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,15]]},"reference":[{"key":"40_CR1","unstructured":"Radford, A., Kim, J.W., Hallacy, C., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning. PMLR, pp. 8748\u20138763 (2021)"},{"issue":"9","key":"40_CR2","doi-asserted-by":"publisher","first-page":"2337","DOI":"10.1007\/s11263-022-01653-1","volume":"130","author":"K Zhou","year":"2022","unstructured":"Zhou, K., Yang, J., Loy, C.C., et al.: Learning to prompt for vision-language models. Int. J. Comput. Vision 130(9), 2337\u20132348 (2022)","journal-title":"Int. J. Comput. Vision"},{"key":"40_CR3","doi-asserted-by":"crossref","unstructured":"Zhou, K., Yang, J., Loy, C.C., et al.: conditional prompt learning for vision-language models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16816\u201316825 (2022)","DOI":"10.1109\/CVPR52688.2022.01631"},{"key":"40_CR4","doi-asserted-by":"crossref","unstructured":"Pratt, S., Covert, I., Liu, R., et al.: What does a platypus look like? generating customized prompts for zero-shot image classification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15691\u201315701 (2023)","DOI":"10.1109\/ICCV51070.2023.01438"},{"issue":"10","key":"40_CR5","doi-asserted-by":"publisher","first-page":"268","DOI":"10.1007\/s10462-024-10915-y","volume":"57","author":"F Liu","year":"2024","unstructured":"Liu, F., Zhang, T., Dai, W., et al.: Few-shot adaptation of multi-modal foundation models: a survey. Artif. Intell. Rev. 57(10), 268 (2024)","journal-title":"Artif. Intell. Rev."},{"issue":"2","key":"40_CR6","doi-asserted-by":"publisher","first-page":"581","DOI":"10.1007\/s11263-023-01891-x","volume":"132","author":"P Gao","year":"2024","unstructured":"Gao, P., Geng, S., Zhang, R., et al.: Clip-Adapter: better vision-language models with feature adapters. Int. J. Comput. Vision 132(2), 581\u2013595 (2024)","journal-title":"Int. J. Comput. Vision"},{"issue":"2","key":"40_CR7","first-page":"3","volume":"1","author":"EJ Hu","year":"2022","unstructured":"Hu, E.J., Shen, Y., Wallis, P., et al.: LoRA: low-rank adaptation of large language models. ICLR. 1(2), 3 (2022)","journal-title":"ICLR."},{"key":"40_CR8","unstructured":"Liu, A., Feng, B., Xue, B., et al.: DeepSeek-v3 Technical Report. arXiv preprint arXiv:2412.19437 (2024)"},{"key":"40_CR9","doi-asserted-by":"crossref","unstructured":"Udandarao, V., Gupta, A., Albanie, S.: SUS-X: Training-free name-only transfer of vision-language models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2725\u20132736 (2023)","DOI":"10.1109\/ICCV51070.2023.00257"},{"key":"40_CR10","doi-asserted-by":"crossref","unstructured":"Zhang, E., Zhu, B., Chen, Y., et al.: Optimization of Prompt Learning via Multi-Knowledge Representation for Vision-Language Models. arXiv preprint arXiv:2404.10357 (2024)","DOI":"10.1109\/TMM.2025.3599096"},{"key":"40_CR11","unstructured":"Khattak, M.U., Naeem, M.F., Naseer, M., et al.: Learning to Prompt with Text Only Supervision for Vision-Language Models. arXiv preprint arXiv:2401.02418 (2024)"},{"key":"40_CR12","doi-asserted-by":"crossref","unstructured":"Roth, K., Kim, J.M., Koepke, A., et al.: waffling around for performance: visual classification with random words and broad concepts. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15746\u201315757 (2023)","DOI":"10.1109\/ICCV51070.2023.01443"},{"key":"40_CR13","doi-asserted-by":"crossref","unstructured":"Sun, Z., Fang, Y., Wu, T., et al.: Alpha-CLIP: A CLIP model focusing on wherever you want. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13019\u201313029 (2024)","DOI":"10.1109\/CVPR52733.2024.01237"},{"key":"40_CR14","doi-asserted-by":"crossref","unstructured":"Khattak, M.U., Rasheed, H., Maaz, M., et al.: MAPLE: multi-modal prompt learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 19113\u201319122 (2023)","DOI":"10.1109\/CVPR52729.2023.01832"},{"key":"40_CR15","doi-asserted-by":"crossref","unstructured":"Zhang, R., Zhang, W., Fang, R., et al.: Tip-adapter: training-free adaption of CLIP for few-shot classification. In: European Conference on Computer Vision. Cham: Springer Nature Switzerland, pp. 493\u2013510 (2022)","DOI":"10.1007\/978-3-031-19833-5_29"},{"key":"40_CR16","doi-asserted-by":"crossref","unstructured":"Maniparambil, M., Vorster, C., Molloy, D., et al.: Enhancing CLIP with GPT-4: harnessing visual descriptions as prompts. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 262\u2013271 (2023)","DOI":"10.1109\/ICCVW60793.2023.00034"},{"key":"40_CR17","unstructured":"Guo, Y., Zhuang, S., Li, K., et al.: TransAgent: transfer vision-language foundation models with heterogeneous agent collaboration. In: The Thirty-Eighth Annual Conference on Neural Information Processing Systems (2024)"},{"key":"40_CR18","unstructured":"Ramesh, A., Pavlov, M., Goh, G., et al.: Zero-shot text-to-image generation. In: International Conference on Machine Learning. PMLR, pp. 8821\u20138831 (2021)"},{"issue":"4","key":"40_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530068","volume":"41","author":"Y Vinker","year":"2022","unstructured":"Vinker, Y., Pajouheshgar, E., Bo, J.Y., et al.: Clipasso: semantically-aware object sketching. ACM Trans. Graph. (TOG). 41(4), 1\u201311 (2022)","journal-title":"ACM Trans. Graph. (TOG)."},{"key":"40_CR20","doi-asserted-by":"crossref","unstructured":"Zhong, Y., Yang, J., Zhang, P., et al.: RegionCLIP: region-based language-image pretraining. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16793\u201316803 (2022)","DOI":"10.1109\/CVPR52688.2022.01629"},{"key":"40_CR21","doi-asserted-by":"crossref","unstructured":"Li, L.H., Zhang, P., Zhang, H., et al.: Grounded language-image pre-training. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10965\u201310975 (2022)","DOI":"10.1109\/CVPR52688.2022.01069"},{"key":"40_CR22","unstructured":"Li, F.-F.: Learning generative visual models from few training examples. In: Workshop on Generative-Model Based Vision, IEEE Proc. CVPR (2004)"},{"key":"40_CR23","unstructured":"Griffin, G., Holub, A., Perona, P.: Caltech-256 Object Category Dataset. Technical Report 7694, California Institute of Technology, Pasadena (2007)"},{"key":"40_CR24","doi-asserted-by":"crossref","unstructured":"Cimpoi, M., et al.: Describing textures in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2014)","DOI":"10.1109\/CVPR.2014.461"},{"key":"40_CR25","doi-asserted-by":"crossref","unstructured":"Parkhi, O. M., et al.: Cats and Dogs. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6248092"},{"key":"40_CR26","unstructured":"Wah, C., et al.: The Caltech-UCSD Birds-200\u20132011 Dataset (2011)"},{"key":"40_CR27","doi-asserted-by":"crossref","unstructured":"Bossard, L., Guillaumin, M., Van Gool, L.: Food-101 \u2013 mining discriminative components with random forests. In: Computer Vision \u2013 ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part VI. Springer International Publishing, pp. (2014)","DOI":"10.1007\/978-3-319-10599-4_29"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-9794-6_40","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T07:57:48Z","timestamp":1757231868000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-9794-6_40"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819697939","9789819697946"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-9794-6_40","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"15 July 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ningbo","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 July 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 July 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/icg\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}