{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,6,26]],"date-time":"2024-06-26T00:14:12Z","timestamp":1719360852640},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"S1","license":[{"start":{"date-parts":[[2024,4,18]],"date-time":"2024-04-18T00:00:00Z","timestamp":1713398400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,18]],"date-time":"2024-04-18T00:00:00Z","timestamp":1713398400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s11760-024-03200-6","type":"journal-article","created":{"date-parts":[[2024,4,18]],"date-time":"2024-04-18T04:01:31Z","timestamp":1713412891000},"page":"883-893","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Source bias reduction for source-free domain adaptation"],"prefix":"10.1007","volume":"18","author":[{"given":"Liang","family":"Tian","sequence":"first","affiliation":[]},{"given":"Mao","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Lihua","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Zhenbin","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,18]]},"reference":[{"key":"3200_CR1","unstructured":"Ganin, Y., Lempitsky, V.: Unsupervised domain adaptation by backpropagation, In: International Conference on Machine Learning, pp. 1180\u20131189. PMLR (2015)"},{"key":"3200_CR2","doi-asserted-by":"crossref","unstructured":"Saito, K., Watanabe, K., Ushiku, Y., Harada, T.: Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3723\u20133732 (2018)","DOI":"10.1109\/CVPR.2018.00392"},{"key":"3200_CR3","unstructured":"Long, M., Cao, Z., Wang, J., Jordan, M.I.: Conditional adversarial domain adaptation. In: Advances in Neural Information Processing Systems, vol. 31 (2018)"},{"key":"3200_CR4","doi-asserted-by":"publisher","first-page":"110246","DOI":"10.1016\/j.patcog.2023.110246","volume":"149","author":"J He","year":"2024","unstructured":"He, J., Wu, L., Tao, C., Lv, F.: Source-free domain adaptation with unrestricted source hypothesis. Pattern Recognit. 149, 110246 (2024)","journal-title":"Pattern Recognit."},{"key":"3200_CR5","unstructured":"Yang, S., Wang, Y., Van De Weijer, J., Herranz, L., Jui, S.: Unsupervised domain adaptation without source data by casting a bait, 1(2), 5 (2020). arXiv:2010.12427"},{"key":"3200_CR6","doi-asserted-by":"crossref","unstructured":"Qiu, Z., Zhang, Y., Lin, H., Niu, S., Liu, Y., Du, Q., Tan, M.: Source-free domain adaptation via avatar prototype generation and adaptation. arXiv:2106.15326 (2021)","DOI":"10.24963\/ijcai.2021\/402"},{"key":"3200_CR7","unstructured":"Liang, J., Hu, D., Feng, J.: Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In: International Conference on Machine Learning, pp. 6028\u20136039. PMLR (2020)"},{"key":"3200_CR8","doi-asserted-by":"crossref","unstructured":"Yang, S., Wang, Y., Van De Weijer, J., Herranz, L., Jui, S.: Generalized source-free domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8978\u20138987 (2021)","DOI":"10.1109\/ICCV48922.2021.00885"},{"issue":"5","key":"3200_CR9","doi-asserted-by":"publisher","first-page":"2405","DOI":"10.1007\/s11760-022-02457-z","volume":"17","author":"L Tian","year":"2023","unstructured":"Tian, L., Zhou, L., Zhang, H., Wang, Z., Ye, M.: Robust self-supervised learning for source-free domain adaptation. Signal Image Video Process. 17(5), 2405\u20132413 (2023)","journal-title":"Signal Image Video Process."},{"key":"3200_CR10","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"3200_CR11","unstructured":"Ge, C., Huang, R., Xie, M., Lai, Z., Song, S., Li, S., Huang, G.: Domain adaptation via prompt learning. arXiv:2202.06687 (2022)"},{"key":"3200_CR12","doi-asserted-by":"crossref","unstructured":"Lai, Z., Vesdapunt, N., Zhou, N., Wu, J., Huynh, C. P., Li, X., Fu, K.K., Chuah, C.-N.: PADCLIP: Pseudo-labeling with adaptive debiasing in clip for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 16155\u201316165 (2023)","DOI":"10.1109\/ICCV51070.2023.01480"},{"key":"3200_CR13","unstructured":"Zellinger, W., Grubinger, T., Lughofer, E., Natschl\u00e4ger, T., Saminger-Platz, S.: Central moment discrepancy (CMD) for domain-invariant representation learning. arXiv:1702.08811 (2017)"},{"key":"3200_CR14","doi-asserted-by":"crossref","unstructured":"Kang, G., Jiang, L., Yang, Y., Hauptmann, A.G.: Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4893\u20134902 (2019)","DOI":"10.1109\/CVPR.2019.00503"},{"key":"3200_CR15","doi-asserted-by":"crossref","unstructured":"Li, M., Zhai, Y.-M., Luo, Y.-W., Ge, P.-F., Ren, C.-X.: Enhanced transport distance for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision And Pattern Recognition, pp. 13936\u201313944 (2020)","DOI":"10.1109\/CVPR42600.2020.01395"},{"key":"3200_CR16","unstructured":"Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., Lempitsky, V.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(1), 2096\u20132030 (2016)"},{"key":"3200_CR17","doi-asserted-by":"crossref","unstructured":"Tzeng, E., Hoffman, J., Saenko, K., Darrell, T.: Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7167\u20137176 (2017)","DOI":"10.1109\/CVPR.2017.316"},{"key":"3200_CR18","unstructured":"Liu, M.-Y., Tuzel, O.: Coupled generative adversarial networks. In: Advances in Neural Information Processing Systems, vol. 29 (2016)"},{"key":"3200_CR19","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"3200_CR20","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2024.3354420","author":"M Wang","year":"2024","unstructured":"Wang, M., Liu, Y., Yuan, J., Wang, S., Wang, Z., Wang, W.: Inter-class and inter-domain semantic augmentation for domain generalization. IEEE Trans. Image Process. (2024). https:\/\/doi.org\/10.1109\/TIP.2024.3354420","journal-title":"IEEE Trans. Image Process."},{"key":"3200_CR21","doi-asserted-by":"crossref","unstructured":"Ghifary, M., Kleijn, W. B., Zhang, M., Balduzzi, D.: Domain generalization for object recognition with multi-task autoencoders. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2551\u20132559 (2015)","DOI":"10.1109\/ICCV.2015.293"},{"key":"3200_CR22","unstructured":"Sun, Y., Tzeng, E., Darrell, T., Efros, A. A.: Unsupervised domain adaptation through self-supervision. arXiv:1909.11825 (2019)"},{"issue":"6","key":"3200_CR23","doi-asserted-by":"publisher","first-page":"3749","DOI":"10.1109\/TCSVT.2021.3111034","volume":"32","author":"J Tian","year":"2021","unstructured":"Tian, J., Zhang, J., Li, W., Xu, D.: VDM-DA: virtual domain modeling for source data-free domain adaptation. IEEE Trans. Circuits Syst. Video Technol. 32(6), 3749\u20133760 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3200_CR24","doi-asserted-by":"crossref","unstructured":"Quattoni, A., Collins, M., Darrell, T.: Learning visual representations using images with captions. In: 2007 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1\u20138. IEEE (2007)","DOI":"10.1109\/CVPR.2007.383173"},{"key":"3200_CR25","unstructured":"Srivastava, N., Salakhutdinov, R.R.: Multimodal learning with deep Boltzmann machines. In: Advances in Neural Information Processing Systems, vol. 25 (2012)"},{"key":"3200_CR26","unstructured":"Jia, C., Yang, Y., Xia, Y., Chen, Y.-T., Parekh, Z., Pham, H., Le, Q., Sung, Y.-H., Li, Z., Duerig, T.: Scaling up visual and vision-language representation learning with noisy text supervision. In: International Conference on Machine Learning, pp. 4904\u20134916. PMLR (2021)"},{"issue":"4","key":"3200_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530164","volume":"41","author":"R Gal","year":"2022","unstructured":"Gal, R., Patashnik, O., Maron, H., Bermano, A.H., Chechik, G., Cohen-Or, D.: StyGAN-NADA: CLIP-guided domain adaptation of image generators. ACM Trans. Actions Gr. (TOG) 41(4), 1\u201313 (2022)","journal-title":"ACM Trans. Actions Gr. (TOG)"},{"key":"3200_CR28","doi-asserted-by":"crossref","unstructured":"Vidit, V., Engilberge, M., Salzmann, M.: Clip the gap: a single domain generalization approach for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3219\u20133229 (2023)","DOI":"10.1109\/CVPR52729.2023.00314"},{"key":"3200_CR29","doi-asserted-by":"crossref","unstructured":"Fahes, M., Vu, T.-H., Bursuc, A., P\u00e9rez, P., de Charette, R.: PODA: prompt-driven zero-shot domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 18623\u201318633 (2023)","DOI":"10.1109\/ICCV51070.2023.01707"},{"key":"3200_CR30","doi-asserted-by":"crossref","unstructured":"Khattak, M.U., Wasim, S.T., Naseer, M., Khan, S., Yang, M.-H., Khan, F.S.: Self-regulating prompts: foundational model adaptation without forgetting. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15190\u201315200 (2023)","DOI":"10.1109\/ICCV51070.2023.01394"},{"key":"3200_CR31","doi-asserted-by":"crossref","unstructured":"Zara, G., Roy, S., Rota, P., Ricci, E.: AutoLabel: CLIP-based framework for open-set video domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11504\u201311513 (2023)","DOI":"10.1109\/CVPR52729.2023.01107"},{"issue":"6","key":"3200_CR32","doi-asserted-by":"publisher","first-page":"508","DOI":"10.1109\/TAI.2021.3110179","volume":"2","author":"Y Kim","year":"2021","unstructured":"Kim, Y., Cho, D., Han, K., Panda, P., Hong, S.: Domain adaptation without source data. IEEE Trans. Artif. Intell. 2(6), 508\u2013518 (2021)","journal-title":"IEEE Trans. Artif. Intell."},{"issue":"1","key":"3200_CR33","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1214\/aoms\/1177729694","volume":"22","author":"S Kullback","year":"1951","unstructured":"Kullback, S., Leibler, R.A.: On information and sufficiency. Ann. Math. Stat. 22(1), 79\u201386 (1951)","journal-title":"Ann. Math. Stat."},{"key":"3200_CR34","doi-asserted-by":"crossref","unstructured":"Saenko, K., Kulis, B., Fritz, M., Darrell, T.: Adapting visual category models to new domains. In: European Conference on Computer Vision, pp. 213\u2013226. Springer (2010)","DOI":"10.1007\/978-3-642-15561-1_16"},{"key":"3200_CR35","doi-asserted-by":"crossref","unstructured":"Venkateswara, H., Eusebio, J., Chakraborty, S., Panchanathan, S.: Deep hashing network for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5018\u20135027 (2017)","DOI":"10.1109\/CVPR.2017.572"},{"key":"3200_CR36","unstructured":"Peng, X., Usman, B., Kaushik, N., Hoffman, J., Wang, D., Saenko, K.: VisDA: the visual domain adaptation challenge. arXiv:1710.06924 (2017)"},{"key":"3200_CR37","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3200_CR38","doi-asserted-by":"crossref","unstructured":"Wu, Y., Inkpen, D., El-Roby, A.: Dual mixup regularized learning for adversarial domain adaptation. In: European Conference on Computer Vision, pp. 540\u2013555. Springer (2020)","DOI":"10.1007\/978-3-030-58526-6_32"},{"key":"3200_CR39","unstructured":"Chen, X., Wang, S., Long, M., Wang, J.: Transferability versus discriminability: batch spectral penalization for adversarial domain adaptation. In: International Conference on Machine Learning, PMLR, pp. 1081\u20131090 (2019)"},{"key":"3200_CR40","doi-asserted-by":"crossref","unstructured":"Xu, R., Li, G., Yang, J., Lin, L.: Larger norm more transferable: an adaptive feature norm approach for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1426\u20131435 (2019)","DOI":"10.1109\/ICCV.2019.00151"},{"key":"3200_CR41","doi-asserted-by":"crossref","unstructured":"Cui, S., Wang, S., Zhuo, J., Li, L., Huang, Q., Tian, Q.: Towards discriminability and diversity: batch nuclear-norm maximization under label insufficient situations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3941\u20133950 (2020)","DOI":"10.1109\/CVPR42600.2020.00400"},{"key":"3200_CR42","doi-asserted-by":"crossref","unstructured":"Yang, G., Xia, H., Ding, M., Ding, Z.: Bi-directional generation for unsupervised domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 6615\u20136622 (2020)","DOI":"10.1609\/aaai.v34i04.6137"},{"key":"3200_CR43","doi-asserted-by":"crossref","unstructured":"Jin, Y., Wang, X., Long, M., Wang, J.: Minimum class confusion for versatile domain adaptation. In: European Conference on Computer Vision, pp. 464\u2013480. Springer (2020)","DOI":"10.1007\/978-3-030-58589-1_28"},{"key":"3200_CR44","doi-asserted-by":"crossref","unstructured":"Tang, H., Chen, K., Jia, K.: Unsupervised domain adaptation via structurally regularized deep clustering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8725\u20138735, (2020)","DOI":"10.1109\/CVPR42600.2020.00875"},{"key":"3200_CR45","doi-asserted-by":"crossref","unstructured":"Liang, J., Hu, D., Feng, J.: Domain adaptation with auxiliary target domain-oriented classifier. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16632\u201316642 (2021)","DOI":"10.1109\/CVPR46437.2021.01636"},{"key":"3200_CR46","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2023.3346444","author":"M Wang","year":"2024","unstructured":"Wang, M., Wang, S., Yang, X., Yuan, J., Zhang, W.: Equity in unsupervised domain adaptation by nuclear norm maximization. IEEE Trans. Circuits Syst. Video Technol. (2024). https:\/\/doi.org\/10.1109\/TCSVT.2023.3346444","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3200_CR47","doi-asserted-by":"crossref","unstructured":"Li, R., Jiao, Q., Cao, W., Wong, H.-S., Wu, S.: Model adaptation: unsupervised domain adaptation without source data, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9641\u20139650 (2020)","DOI":"10.1109\/CVPR42600.2020.00966"},{"key":"3200_CR48","first-page":"29393","volume":"34","author":"S Yang","year":"2021","unstructured":"Yang, S., van de Weijer, J., Herranz, L., Jui, S., et al.: Exploiting the intrinsic neighborhood structure for source-free domain adaptation. Adv. Neural Inf. Process. Syst. 34, 29393\u201329405 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3200_CR49","doi-asserted-by":"publisher","first-page":"467","DOI":"10.1016\/j.neunet.2022.05.015","volume":"152","author":"S Tang","year":"2022","unstructured":"Tang, S., Zou, Y., Song, Z., Lyu, J., Chen, L., Ye, M., Zhong, S., Zhang, J.: Semantic consistency learning on manifold for source data-free unsupervised domain adaptation. Neural Netw. 152, 467\u2013478 (2022)","journal-title":"Neural Netw."},{"key":"3200_CR50","doi-asserted-by":"crossref","unstructured":"Ding, Y., Sheng, L., Liang, J., Zheng, A., He, R.: Proxymix: Proxy-based mixup training with label refinery for source-free domain adaptation, arXiv:2205.14566 (2022)","DOI":"10.1016\/j.neunet.2023.08.005"},{"key":"3200_CR51","doi-asserted-by":"publisher","DOI":"10.1049\/cit2.12228","author":"S Tang","year":"2023","unstructured":"Tang, S., Su, W., Yang, Y., Chen, L., Ye, M.: Model adaptation via credible local context representation. CAAI Trans. Intell. Technol. (2023). https:\/\/doi.org\/10.1049\/cit2.12228","journal-title":"CAAI Trans. Intell. Technol."},{"issue":"11","key":"3200_CR52","first-page":"2579","volume":"9","author":"L Van der Maaten","year":"2008","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11), 2579\u20132605 (2008)","journal-title":"J. Mach. Learn. Res."}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03200-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03200-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03200-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,25]],"date-time":"2024-06-25T12:24:00Z","timestamp":1719318240000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03200-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,18]]},"references-count":52,"journal-issue":{"issue":"S1","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["3200"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03200-6","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4,18]]},"assertion":[{"value":"18 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 March 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 April 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 April 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This declaration is not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This declaration is not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}