{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T16:39:06Z","timestamp":1769186346996,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":42,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556816","type":"print"},{"value":"9789819556823","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5682-3_15","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:55Z","timestamp":1769116435000},"page":"209-224","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multimodal Adapter-Driven Source-Free Domain Adaptation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3824-687X","authenticated-orcid":false,"given":"Shanshan","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-1475-752X","authenticated-orcid":false,"given":"Houmeng","family":"He","sequence":"additional","affiliation":[]},{"given":"Keyang","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0201-1638","authenticated-orcid":false,"given":"Xun","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"key":"15_CR1","doi-asserted-by":"crossref","unstructured":"Chen, D., Wang, D., Darrell, T., Ebrahimi, S.: Contrastive test-time adaptation. In: Proceedings of the IEEE\/CVF CVPR, pp. 295\u2013305 (2022)","DOI":"10.1109\/CVPR52688.2022.00039"},{"key":"15_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y., Wang, Y., Pan, Y., Yao, T., Tian, X., Mei, T.: A style and semantic memory mechanism for domain generalization. In: Proceedings of the IEEE\/CVF ICCV, pp. 9164\u20139173 (2021)","DOI":"10.1109\/ICCV48922.2021.00903"},{"key":"15_CR3","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"issue":"6","key":"15_CR4","doi-asserted-by":"publisher","first-page":"3611","DOI":"10.1007\/s10994-023-06432-8","volume":"113","author":"Y Du","year":"2024","unstructured":"Du, Y., et al.: Generation, augmentation, and alignment: a pseudo-source domain based method for source-free domain adaptation. Mach. Learn. 113(6), 3611\u20133631 (2024)","journal-title":"Mach. Learn."},{"key":"15_CR5","unstructured":"Ge, C., Huang, R., Xie, M., Lai, Z., Song, S., Li, S., Huang, G.: Domain adaptation via prompt learning. IEEE Tnnls (2023)"},{"key":"15_CR6","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"15_CR7","first-page":"3635","volume":"34","author":"J Huang","year":"2021","unstructured":"Huang, J., Guan, D., Xiao, A., Lu, S.: Model adaptation: historical contrastive learning for unsupervised domain adaptation without source data. NeurIPS 34, 3635\u20133649 (2021)","journal-title":"NeurIPS"},{"key":"15_CR8","unstructured":"Jia, C., et al.: Scaling up visual and vision-language representation learning with noisy text supervision. In: ICML, pp. 4904\u20134916. PMLR (2021)"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Kang, J., Lee, S., Kim, N., Kwak, S.: Style neophile: constantly seeking novel styles for domain generalization. In: Proceedings of the IEEE\/CVF CVPR, pp. 7130\u20137140 (2022)","DOI":"10.1109\/CVPR52688.2022.00699"},{"key":"15_CR10","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NeurIPS, vol. 25 (2012)"},{"key":"15_CR11","doi-asserted-by":"crossref","unstructured":"Lao, Q., Jiang, X., Havaei, M.: Hypothesis disparity regularized mutual information maximization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a035, pp. 8243\u20138251 (2021)","DOI":"10.1609\/aaai.v35i9.17003"},{"key":"15_CR12","unstructured":"Lee, J., Jung, D., Yim, J., Yoon, S.: Confidence score for source-free unsupervised domain adaptation. In: ICML, pp. 12365\u201312377. PMLR (2022)"},{"key":"15_CR13","doi-asserted-by":"crossref","unstructured":"Li, H., Pan, S.J., Wang, S., Kot, A.C.: Domain generalization with adversarial feature learning. In: Proceedings of the IEEE CVPR, pp. 5400\u20135409 (2018)","DOI":"10.1109\/CVPR.2018.00566"},{"key":"15_CR14","doi-asserted-by":"crossref","unstructured":"Liang, J., He, R., Sun, Z., Tan, T.: Distant supervised centroid shift: a simple and efficient approach to visual domain adaptation. In: Proceedings of the IEEE\/CVF CVPR, pp. 2975\u20132984 (2019)","DOI":"10.1109\/CVPR.2019.00309"},{"key":"15_CR15","unstructured":"Liang, J., Hu, D., Feng, J.: Do we really need to access the source data? Source hypothesis transfer for unsupervised domain adaptation. In: ICML, pp. 6028\u20136039. PMLR (2020)"},{"key":"15_CR16","doi-asserted-by":"crossref","unstructured":"Litrico, M., Del\u00a0Bue, A., Morerio, P.: Guiding pseudo-labels with uncertainty estimation for source-free unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF CVPR, pp. 7640\u20137650 (2023)","DOI":"10.1109\/CVPR52729.2023.00738"},{"key":"15_CR17","doi-asserted-by":"crossref","unstructured":"Lv, F., et al.: Causality inspired representation learning for domain generalization. In: Proceedings of the IEEE\/CVF CVPR, pp. 8046\u20138056 (2022)","DOI":"10.1109\/CVPR52688.2022.00788"},{"key":"15_CR18","doi-asserted-by":"crossref","unstructured":"Peng, X., Bai, Q., Xia, X., Huang, Z., Saenko, K., Wang, B.: Moment matching for multi-source domain adaptation. In: Proceedings of the IEEE\/CVF ICCV, pp. 1406\u20131415 (2019)","DOI":"10.1109\/ICCV.2019.00149"},{"key":"15_CR19","unstructured":"Peng, X., Usman, B., Kaushik, N., Hoffman, J., Wang, D., Saenko, K.: Visda: the visual domain adaptation challenge. arXiv preprint arXiv:1710.06924 (2017)"},{"key":"15_CR20","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: ICML, pp. 8748\u20138763. PmLR (2021)"},{"key":"15_CR21","doi-asserted-by":"publisher","unstructured":"Saenko, K., Kulis, B., Fritz, M., Darrell, T.: Adapting visual category models to new domains. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6314, pp. 213\u2013226. Springer, Heidelberg (2010). https:\/\/doi.org\/10.1007\/978-3-642-15561-1_16","DOI":"10.1007\/978-3-642-15561-1_16"},{"key":"15_CR22","doi-asserted-by":"crossref","unstructured":"Singha, M., Pal, H., Jha, A., Banerjee, B.: Ad-clip: adapting domains in prompt space using clip. In: Proceedings of the IEEE\/CVF ICCV, pp. 4355\u20134364 (2023)","DOI":"10.1109\/ICCVW60793.2023.00470"},{"key":"15_CR23","doi-asserted-by":"publisher","unstructured":"Sun, B., Saenko, K.: Deep CORAL: correlation alignment for deep domain adaptation. In: Hua, G., J\u00e9gou, H. (eds.) ECCV 2016. LNCS, vol. 9915, pp. 443\u2013450. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-49409-8_35","DOI":"10.1007\/978-3-319-49409-8_35"},{"issue":"3","key":"15_CR24","doi-asserted-by":"publisher","first-page":"654","DOI":"10.1007\/s11263-023-01892-w","volume":"132","author":"S Tang","year":"2024","unstructured":"Tang, S., Chang, A., Zhang, F., Zhu, X., Ye, M., Zhang, C.: Source-free domain adaptation via target prediction distribution searching. IJCV 132(3), 654\u2013672 (2024)","journal-title":"IJCV"},{"key":"15_CR25","doi-asserted-by":"crossref","unstructured":"Tang, S., et al.: Model adaptation through hypothesis transfer with gradual knowledge distillation. In: 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5679\u20135685. IEEE (2021)","DOI":"10.1109\/IROS51168.2021.9636206"},{"key":"15_CR26","doi-asserted-by":"crossref","unstructured":"Tang, S., Su, W., Ye, M., Zhu, X.: Source-free domain adaptation with frozen multimodal foundation model. In: Proceedings of the IEEE\/CVF CVPR, pp. 23711\u201323720 (2024)","DOI":"10.1109\/CVPR52733.2024.02238"},{"issue":"6","key":"15_CR27","doi-asserted-by":"publisher","first-page":"3749","DOI":"10.1109\/TCSVT.2021.3111034","volume":"32","author":"J Tian","year":"2021","unstructured":"Tian, J., Zhang, J., Li, W., Xu, D.: Vdm-da: virtual domain modeling for source data-free domain adaptation. IEEE Trans. Circuits Syst. Video Technol. 32(6), 3749\u20133760 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"15_CR28","doi-asserted-by":"crossref","unstructured":"Venkateswara, H., Eusebio, J., Chakraborty, S., Panchanathan, S.: Deep hashing network for unsupervised domain adaptation. In: Proceedings of the IEEE CVPR, pp. 5018\u20135027 (2017)","DOI":"10.1109\/CVPR.2017.572"},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Wang, S., ALuSi, Yang, X., Xu, K., Tan, H., Zhang, X.: Dual-stream feature augmentation for domain generalization. In: Proceedings of the 32nd ACM International Conference on Multimedia, pp. 1111\u20131119 (2024)","DOI":"10.1145\/3664647.3680652"},{"key":"15_CR30","doi-asserted-by":"crossref","unstructured":"Wang, S., et al.: Disentangled representation learning with causality for unsupervised domain adaptation. In: Proceedings of the 31st ACM International Conference on Multimedia, pp. 2918\u20132926 (2023)","DOI":"10.1145\/3581783.3611725"},{"key":"15_CR31","doi-asserted-by":"crossref","unstructured":"Wang, S., et al.: Exploring invariance matters for domain generalization. IEEE Trans. Image Process. (2025)","DOI":"10.1109\/TIP.2025.3568747"},{"key":"15_CR32","doi-asserted-by":"crossref","unstructured":"Wang, S., Shen, X., Yang, X., Xu, K., Zhang, X.: Feature responsive lora: towards parameter-efficient transfer learning for self-supervised visual models. IEEE Trans. Circ. Syst. Video Technol. (2025)","DOI":"10.1109\/TCSVT.2025.3595896"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Wang, S., Zhang, L.: Self-adaptive re-weighted adversarial domain adaptation. arXiv preprint arXiv:2006.00223 (2020)","DOI":"10.24963\/ijcai.2020\/440"},{"key":"15_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.108993","volume":"133","author":"S Wang","year":"2023","unstructured":"Wang, S., Zhang, L., Wang, P., Wang, M., Zhang, X.: Bp-triplet net for unsupervised domain adaptation: a bayesian perspective. Pattern Recogn. 133, 108993 (2023)","journal-title":"Pattern Recogn."},{"key":"15_CR35","doi-asserted-by":"publisher","first-page":"2424","DOI":"10.1109\/TIP.2019.2948480","volume":"29","author":"S Wang","year":"2019","unstructured":"Wang, S., Zhang, L., Zuo, W., Zhang, B.: Class-specific reconstruction transfer learning for visual recognition across domains. IEEE Trans. Image Process. 29, 2424\u20132438 (2019)","journal-title":"IEEE Trans. Image Process."},{"issue":"8","key":"15_CR36","doi-asserted-by":"publisher","first-page":"5495","DOI":"10.1109\/TCSVT.2022.3152615","volume":"32","author":"Y Wang","year":"2022","unstructured":"Wang, Y., Qi, L., Shi, Y., Gao, Y.: Feature-based style randomization for domain generalization. IEEE Trans. Circ. Syst. Video Technol. 32(8), 5495\u20135509 (2022)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"15_CR37","doi-asserted-by":"crossref","unstructured":"Wang, Z., Loog, M., Van\u00a0Gemert, J.: Respecting domain relations: hypothesis invariance for domain generalization. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 9756\u20139763. IEEE (2021)","DOI":"10.1109\/ICPR48806.2021.9412797"},{"key":"15_CR38","doi-asserted-by":"crossref","unstructured":"Xu, Q., Zhang, R., Zhang, Y., Wang, Y., Tian, Q.: A Fourier-based framework for domain generalization. In: Proceedings of the IEEE\/CVF CVPR, pp. 14383\u201314392 (2021)","DOI":"10.1109\/CVPR46437.2021.01415"},{"key":"15_CR39","unstructured":"Yan, S., Song, H., Li, N., Zou, L., Ren, L.: Improve unsupervised domain adaptation with mixup training. arXiv preprint arXiv:2001.00677 (2020)"},{"key":"15_CR40","first-page":"29393","volume":"34","author":"S Yang","year":"2021","unstructured":"Yang, S., Van de Weijer, J., Herranz, L., Jui, S., et al.: Exploiting the intrinsic neighborhood structure for source-free domain adaptation. NeurIPS 34, 29393\u201329405 (2021)","journal-title":"NeurIPS"},{"key":"15_CR41","doi-asserted-by":"crossref","unstructured":"Zhang, W., Shen, L., Foo, C.S.: Source-free domain adaptation guided by vision and vision-language pre-training. IJCV 1\u201323 (2024)","DOI":"10.1007\/s11263-024-02215-3"},{"key":"15_CR42","doi-asserted-by":"crossref","unstructured":"Zhou, K., Yang, Y., Hospedales, T., Xiang, T.: Deep domain-adversarial image generation for domain generalisation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a034, pp. 13025\u201313032 (2020)","DOI":"10.1609\/aaai.v34i07.7003"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5682-3_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:14:06Z","timestamp":1769116446000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5682-3_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556816","9789819556823"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5682-3_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}