{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,1,18]],"date-time":"2025-01-18T05:08:27Z","timestamp":1737176907891,"version":"3.33.0"},"reference-count":50,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,12,15]],"date-time":"2024-12-15T00:00:00Z","timestamp":1734220800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,15]],"date-time":"2024-12-15T00:00:00Z","timestamp":1734220800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,12,15]]},"DOI":"10.1109\/bigdata62323.2024.10825383","type":"proceedings-article","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T18:31:23Z","timestamp":1737052283000},"page":"3343-3352","source":"Crossref","is-referenced-by-count":0,"title":["Disentangled Prompt Learning for Transferable, Multimodal, Few-Shot Image Classification"],"prefix":"10.1109","author":[{"given":"John","family":"Yang","sequence":"first","affiliation":[{"name":"Stanford University,Stanford,California,United States"}]},{"given":"Alessandro","family":"Magnani","sequence":"additional","affiliation":[{"name":"Walmart Global Tech,Sunnyvale,California,United States"}]},{"given":"Binwei","family":"Yang","sequence":"additional","affiliation":[{"name":"Walmart Global Tech,Sunnyvale,California,United States"}]}],"member":"263","reference":[{"year":"2021","author":"Radford","article-title":"Learning transferable visual models from natural language supervision","key":"ref1"},{"issue":"9","key":"ref2","first-page":"2337","article-title":"Learning to prompt for vision-language models","volume-title":"International Journal of Computer Vision","volume":"130","author":"Zhou","year":"2022"},{"year":"2022","author":"Liu","article-title":"Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning","key":"ref3"},{"year":"2021","author":"Zhang","article-title":"Vt-clip: Enhancing vision-language models with visual-guided texts","key":"ref4"},{"year":"2022","author":"Li","article-title":"Masked unsupervised self-training for zero-shot image classification","key":"ref5"},{"doi-asserted-by":"publisher","key":"ref6","DOI":"10.1109\/CVPR52729.2023.01832"},{"year":"2021","author":"Andreassen","article-title":"The evolution of out-of-distribution robustness throughout fine-tuning","key":"ref7"},{"year":"2021","author":"Liu","article-title":"Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing","key":"ref8"},{"key":"ref9","doi-asserted-by":"crossref","DOI":"10.1109\/CVPR52688.2022.00514","article-title":"Prompt distribution learning","author":"Lu","year":"2022"},{"key":"ref10","doi-asserted-by":"crossref","DOI":"10.1109\/CVPR52688.2022.01631","article-title":"Conditional prompt learning for vision-language models","author":"Zhou","year":"2022"},{"key":"ref11","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2021.findings-acl.161","article-title":"Template-based named entity recognition using bart","author":"Cui","year":"2021"},{"year":"2019","author":"Jiang","article-title":"How can we know what language models know?","key":"ref12"},{"key":"ref13","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2020.emnlp-main.346","article-title":"Autoprompt: Eliciting knowledge from language models with automatically generated prompts","author":"Shin","year":"2020"},{"year":"2022","author":"Bahng","article-title":"Exploring visual prompts for adapting large-scale models","key":"ref14"},{"key":"ref15","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2021.emnlp-main.243","article-title":"The power of scale for parameter-efficient prompt tuning","author":"Lester","year":"2021"},{"key":"ref16","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/2021.acl-long.353","article-title":"Prefix-tuning: Optimizing continuous prompts for generation","author":"Li","year":"2021"},{"key":"ref17","first-page":"5203","article-title":"Learning how to ask: Querying LMs with mixtures of soft prompts","volume-title":"Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","author":"Qin"},{"key":"ref18","first-page":"14 274","article-title":"Test-time prompt tuning for zero-shot generalization in vision-language models","volume":"35","author":"Shu","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"doi-asserted-by":"publisher","key":"ref19","DOI":"10.1109\/ICCV51070.2023.00257"},{"doi-asserted-by":"publisher","key":"ref20","DOI":"10.1007\/978-3-031-19833-5_7"},{"key":"ref21","first-page":"2763","article-title":"A good prompt is worth millions of parameters: Low-resource prompt-based learning for vision-language models","volume-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Jin"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.1109\/tnnls.2023.3327962"},{"key":"ref23","doi-asserted-by":"crossref","DOI":"10.1109\/CVPR52688.2022.00514","article-title":"Prompt distribution learning","author":"Lu","year":"2022"},{"doi-asserted-by":"publisher","key":"ref24","DOI":"10.1613\/jair.1.11674"},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.1109\/CVPR42600.2020.01133"},{"year":"2023","author":"Nayak","article-title":"Learning to compose soft prompts for compositional zero-shot learning","key":"ref26"},{"doi-asserted-by":"publisher","key":"ref27","DOI":"10.1109\/cvpr52729.2023.02256"},{"year":"2023","author":"Lu","article-title":"Drpt: Disentangled and recurrent prompt tuning for compositional zero-shot learning","key":"ref28"},{"doi-asserted-by":"publisher","key":"ref29","DOI":"10.1007\/978-3-031-72630-9_7"},{"doi-asserted-by":"publisher","key":"ref30","DOI":"10.1109\/CVPR52688.2022.00912"},{"doi-asserted-by":"publisher","key":"ref31","DOI":"10.1109\/CVPR46437.2021.00518"},{"doi-asserted-by":"publisher","key":"ref32","DOI":"10.1109\/CVPR.2015.7298744"},{"doi-asserted-by":"publisher","key":"ref33","DOI":"10.1109\/CVPR46437.2021.00101"},{"doi-asserted-by":"publisher","key":"ref34","DOI":"10.1109\/CVPR.2014.32"},{"doi-asserted-by":"publisher","key":"ref35","DOI":"10.1109\/ICCVW.2019.00377"},{"doi-asserted-by":"publisher","key":"ref36","DOI":"10.18653\/v1\/2022.acl-short.1"},{"year":"2019","author":"Houlsby","article-title":"Parameter-efficient transfer learning for nlp","key":"ref37"},{"key":"ref38","first-page":"4884","article-title":"Parameter-efficient transfer learning with diff pruning","volume-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","author":"Guo"},{"year":"2021","author":"Wortsman","article-title":"Robust fine-tuning of zero-shot models","key":"ref39"},{"doi-asserted-by":"publisher","key":"ref40","DOI":"10.1007\/s11263-023-01891-x"},{"doi-asserted-by":"publisher","key":"ref41","DOI":"10.1007\/978-3-031-19833-5_29"},{"doi-asserted-by":"publisher","key":"ref42","DOI":"10.1109\/tmm.2023.3311646"},{"year":"2023","author":"Chen","article-title":"Plot: Prompt learning with optimal transport for vision-language models","key":"ref43"},{"year":"2022","author":"Huang","article-title":"Unsupervised prompt learning for vision-language models","key":"ref44"},{"year":"2022","author":"Zhu","article-title":"Prompt-aligned gradient for prompt tuning","key":"ref45"},{"doi-asserted-by":"publisher","key":"ref46","DOI":"10.1145\/3447548.3467127"},{"year":"2011","author":"Wah","article-title":"The caltech-ucsd birds-200-2011 dataset","key":"ref47"},{"doi-asserted-by":"publisher","key":"ref48","DOI":"10.1109\/TPAMI.2015.2487986"},{"year":"2021","author":"Ilharco","article-title":"Openclip","key":"ref49"},{"year":"2022","author":"Chia","article-title":"Fashionclip: Connecting language and images for product representations","key":"ref50"}],"event":{"name":"2024 IEEE International Conference on Big Data (BigData)","start":{"date-parts":[[2024,12,15]]},"location":"Washington, DC, USA","end":{"date-parts":[[2024,12,18]]}},"container-title":["2024 IEEE International Conference on Big Data (BigData)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10824975\/10824942\/10825383.pdf?arnumber=10825383","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T08:18:33Z","timestamp":1737101913000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10825383\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,15]]},"references-count":50,"URL":"https:\/\/doi.org\/10.1109\/bigdata62323.2024.10825383","relation":{},"subject":[],"published":{"date-parts":[[2024,12,15]]}}}