{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T12:11:54Z","timestamp":1750075914879,"version":"3.37.3"},"reference-count":72,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,8,16]],"date-time":"2024-08-16T00:00:00Z","timestamp":1723766400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,8,16]],"date-time":"2024-08-16T00:00:00Z","timestamp":1723766400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No.6200609"],"award-info":[{"award-number":["No.6200609"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s00530-024-01444-3","type":"journal-article","created":{"date-parts":[[2024,8,16]],"date-time":"2024-08-16T05:02:17Z","timestamp":1723784537000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Exploiting multi-level consistency learning for source-free domain adaptation"],"prefix":"10.1007","volume":"30","author":[{"given":"Jihong","family":"Ouyang","sequence":"first","affiliation":[]},{"given":"Zhengjie","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Qingyi","family":"Meng","sequence":"additional","affiliation":[]},{"given":"Ximing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jinjin","family":"Chi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,16]]},"reference":[{"key":"1444_CR1","doi-asserted-by":"crossref","unstructured":"Caron, M., Bojanowski, P., Joulin, A. et\u00a0al.: Deep clustering for unsupervised learning of visual features. In: European conference on computer vision, pp. 132\u2013149 (2018)","DOI":"10.1007\/978-3-030-01264-9_9"},{"key":"1444_CR2","doi-asserted-by":"crossref","unstructured":"Chen, C., Xie, W., Huang, W., et\u00a0al.: Progressive feature alignment for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 627\u2013636 (2019)","DOI":"10.1109\/CVPR.2019.00072"},{"key":"1444_CR3","doi-asserted-by":"crossref","unstructured":"Chen, C., Fu, Z., Chen, Z., et\u00a0al.: Homm: Higher-order moment matching for unsupervised domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 3422\u20133429 (2020a)","DOI":"10.1609\/aaai.v34i04.5745"},{"key":"1444_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Q., Du, Y., Tan, Z., et\u00a0al.: Unsupervised domain adaptation with joint domain-adversarial reconstruction networks. In: Machine Learning and Knowledge Discovery in Databases, pp. 640\u2013656 (2021)","DOI":"10.1007\/978-3-030-67661-2_38"},{"key":"1444_CR5","unstructured":"Chen, T., Kornblith, S., Norouzi, M., et\u00a0al.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp. 1597\u20131607 (2020b)"},{"key":"1444_CR6","doi-asserted-by":"crossref","unstructured":"Cicek, S., Soatto, S.: Unsupervised domain adaptation via regularized conditional alignment. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1416\u20131425 (2019)","DOI":"10.1109\/ICCV.2019.00150"},{"key":"1444_CR7","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Mane, D., et\u00a0al.: Autoaugment: learning augmentation strategies from data. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 113\u2013123 (2019)","DOI":"10.1109\/CVPR.2019.00020"},{"key":"1444_CR8","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Shlens, J., et\u00a0al.: (2020) Randaugment: Practical automated data augmentation with a reduced search space. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 702\u2013703","DOI":"10.1109\/CVPRW50498.2020.00359"},{"key":"1444_CR9","unstructured":"French, G., Mackiewicz, M., Fisher, M.: Self-ensembling for visual domain adaptation. In: International Conference on Learning Representations (2018)"},{"issue":"1","key":"1444_CR10","first-page":"2030","volume":"17","author":"Y Ganin","year":"2016","unstructured":"Ganin, Y., Ustinova, E., Ajakan, H., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(1), 2030\u20132096 (2016)","journal-title":"J. Mach. Learn. Res."},{"issue":"3","key":"1444_CR11","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1109\/TCSVT.2021.3075470","volume":"32","author":"J Gao","year":"2021","unstructured":"Gao, J., Xu, C.: Learning video moment retrieval without a single annotated video. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1646\u20131657 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"10","key":"1444_CR12","doi-asserted-by":"publisher","first-page":"3476","DOI":"10.1109\/TPAMI.2020.2985708","volume":"43","author":"J Gao","year":"2020","unstructured":"Gao, J., Zhang, T., Xu, C.: Learning to model relationships for zero-shot video classification. IEEE Trans. Pattern Anal. Mach. Intell. 43(10), 3476\u20133491 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1444_CR13","doi-asserted-by":"crossref","unstructured":"Gao, J., Chen, M., Xu, C.: Vectorized evidential learning for weakly-supervised temporal action localization. IEEE Trans. Pattern Anal. Mach. Intell. (2023)","DOI":"10.1109\/CVPR52729.2023.01416"},{"key":"1444_CR14","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., et\u00a0al.: Generative adversarial nets. In: Neural Information Processing Systems, pp. 2672\u20132680 (2014)"},{"key":"1444_CR15","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.patcog.2023.110246","volume":"149","author":"J He","year":"2024","unstructured":"He, J., Wu, L., Tao, C., et al.: Source-free domain adaptation with unrestricted source hypothesis. Pattern Recogn. 149, 110 (2024)","journal-title":"Pattern Recogn."},{"key":"1444_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., et\u00a0al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016a)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1444_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., et\u00a0al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016b)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1444_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., et\u00a0al.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729\u20139738 (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"1444_CR19","doi-asserted-by":"crossref","unstructured":"Hu, Y., Gao, J., Dong, J., et\u00a0al.: Exploring rich semantics for open-set action recognition. IEEE Trans. Multimed. (2023)","DOI":"10.1109\/TMM.2023.3333206"},{"key":"1444_CR20","unstructured":"Huang, J., Guan, D., Xiao, A., et\u00a0al.: Model adaptation: Historical contrastive learning for unsupervised domain adaptation without source data. In: Neural Information Processing Systems, pp. 3635\u20133649 (2021)"},{"key":"1444_CR21","doi-asserted-by":"crossref","unstructured":"Jin, Y., Wang, X., Long, M., et\u00a0al.: Minimum class confusion for versatile domain adaptation. In: European Conference on Computer Vision, pp. 464\u2013480 (2020)","DOI":"10.1007\/978-3-030-58589-1_28"},{"key":"1444_CR22","doi-asserted-by":"crossref","unstructured":"Kang, G., Jiang, L., Yang, Y., et\u00a0al.: Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4893\u20134902 (2019)","DOI":"10.1109\/CVPR.2019.00503"},{"issue":"6","key":"1444_CR23","doi-asserted-by":"publisher","first-page":"508","DOI":"10.1109\/TAI.2021.3110179","volume":"2","author":"Y Kim","year":"2021","unstructured":"Kim, Y., Cho, D., Han, K., et al.: Domain adaptation without source data. IEEE Trans. Artif. Intell. 2(6), 508\u2013518 (2021)","journal-title":"IEEE Trans. Artif. Intell."},{"key":"1444_CR24","doi-asserted-by":"crossref","unstructured":"Lee, C.Y., Batra, T., Baig, M.H., et\u00a0al.: Sliced wasserstein discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10285\u201310295 (2019)","DOI":"10.1109\/CVPR.2019.01053"},{"key":"1444_CR25","doi-asserted-by":"publisher","first-page":"682","DOI":"10.1016\/j.neunet.2023.02.009","volume":"161","author":"J Lee","year":"2023","unstructured":"Lee, J., Lee, G.: Feature alignment by uncertainty and self-training for source-free unsupervised domain adaptation. Neural Netw. 161, 682\u2013692 (2023). https:\/\/doi.org\/10.1016\/j.neunet.2023.02.009","journal-title":"Neural Netw."},{"issue":"11","key":"1444_CR26","doi-asserted-by":"publisher","first-page":"3918","DOI":"10.1109\/TPAMI.2020.2991050","volume":"43","author":"J Li","year":"2020","unstructured":"Li, J., Chen, E., Ding, Z., et al.: Maximum density divergence for domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 43(11), 3918\u20133930 (2020). https:\/\/doi.org\/10.1109\/TPAMI.2020.2991050","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1444_CR27","doi-asserted-by":"crossref","unstructured":"Li, K., Lu, J., Zuo, H., et\u00a0al.: Source-free multi-domain adaptation with generally auxiliary model training. In: 2022 International Joint Conference on Neural Networks (IJCNN), IEEE, pp. 1\u20138 (2022)","DOI":"10.1109\/IJCNN55064.2022.9892718"},{"key":"1444_CR28","doi-asserted-by":"crossref","unstructured":"Li, K., Lu, J., Zuo, H., et\u00a0al.: Source-free multi-domain adaptation with fuzzy rule-based deep neural networks. IEEE Trans. Fuzzy Syst. (2023)","DOI":"10.1109\/IJCNN55064.2022.9892718"},{"key":"1444_CR29","doi-asserted-by":"crossref","unstructured":"Li, R., Jiao, Q., Cao, W., et\u00a0al.: Model adaptation: Unsupervised domain adaptation without source data. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9641\u20139650 (2020b)","DOI":"10.1109\/CVPR42600.2020.00966"},{"key":"1444_CR30","doi-asserted-by":"crossref","unstructured":"Li, S., Lv, F., Xie, B., et\u00a0al.: Bi-classifier determinacy maximization for unsupervised domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 8455\u20138464 (2021a)","DOI":"10.1609\/aaai.v35i10.17027"},{"key":"1444_CR31","doi-asserted-by":"crossref","unstructured":"Li, S., Xie, M., Gong, K., et\u00a0al.: Transferable semantic augmentation for domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11516\u201311525 (2021b)","DOI":"10.1109\/CVPR46437.2021.01135"},{"key":"1444_CR32","unstructured":"Liang, J., Hu, D., et\u00a0al.: Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In: International Conference on Machine Learning, pp. 6028\u20136039 (2020)"},{"key":"1444_CR33","unstructured":"Liu, H., Wang, J., Long, M.: Cycle self-training for domain adaptation. In: Neural Information Processing Systems, pp. 22968\u201322981 (2021)"},{"key":"1444_CR34","unstructured":"Long, M., Cao, Y., Wang, J., et\u00a0al.: Learning transferable features with deep adaptation networks. In: International Conference on Machine Learning, PMLR, pp. 97\u2013105 (2015)"},{"key":"1444_CR35","unstructured":"Long, M., Zhu, H., Wang, J., et\u00a0al.: Deep transfer learning with joint adaptation networks. In: International Conference on Machine Learning, pp. 2208\u20132217 (2017)"},{"key":"1444_CR36","unstructured":"Long, M., Cao, Z., et\u00a0al.: Conditional adversarial domain adaptation. In: Neural Information Processing Systems, pp. 1647\u20131657 (2018)"},{"key":"1444_CR37","doi-asserted-by":"crossref","unstructured":"Lu, Z., Yang, Y., Zhu, X., et\u00a0al.: Stochastic classifiers for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9111\u20139120 (2020)","DOI":"10.1109\/CVPR42600.2020.00913"},{"key":"1444_CR38","unstructured":"Van\u00a0der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11) (2008)"},{"key":"1444_CR39","doi-asserted-by":"crossref","unstructured":"Mei, K., Zhu, C., Zou, J., et\u00a0al.: Instance adaptive self-training for unsupervised domain adaptation. In: European Conference on Computer Vision, pp. 415\u2013430 (2020)","DOI":"10.1007\/978-3-030-58574-7_25"},{"key":"1444_CR40","doi-asserted-by":"publisher","first-page":"166","DOI":"10.1109\/ACCESS.2021.3136567","volume":"9","author":"BH Ngo","year":"2021","unstructured":"Ngo, B.H., Kim, J.H., Chae, Y.J., et al.: Multi-view collaborative learning for semi-supervised domain adaptation. IEEE Access 9, 166 (2021)","journal-title":"IEEE Access"},{"key":"1444_CR41","doi-asserted-by":"publisher","first-page":"134","DOI":"10.1109\/ACCESS.2023.3337438","volume":"11","author":"BH Ngo","year":"2023","unstructured":"Ngo, B.H., Chae, Y.J., Park, S.J., et al.: Multiple tasks-based multi-source domain adaptation using divide-and-conquer strategy. IEEE Access 11, 134 (2023)","journal-title":"IEEE Access"},{"key":"1444_CR42","doi-asserted-by":"crossref","unstructured":"Pan, F., Shin, I., Rameau, F., et\u00a0al.: Unsupervised intra-domain adaptation for semantic segmentation through self-supervision. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3764\u20133773 (2020)","DOI":"10.1109\/CVPR42600.2020.00382"},{"key":"1444_CR43","first-page":"91","volume":"10","author":"SJ Park","year":"2022","unstructured":"Park, S.J., Park, H.J., Kang, E.S., et al.: Pseudo label rectification via co-teaching and decoupling for multisource domain adaptation in semantic segmentation. IEEE Access 10, 91 (2022)","journal-title":"IEEE Access"},{"key":"1444_CR44","unstructured":"Peng, X., Usman, B., Kaushik, N., et\u00a0al.: Visda: The visual domain adaptation challenge. arXiv preprint arXiv:1710.06924 (2017)"},{"key":"1444_CR45","doi-asserted-by":"crossref","unstructured":"Peng, X., Bai, Q., Xia, X., et\u00a0al.: Moment matching for multi-source domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1406\u20131415 (2019)","DOI":"10.1109\/ICCV.2019.00149"},{"key":"1444_CR46","doi-asserted-by":"crossref","unstructured":"Prabhu, V., Khare, S., Kartik, D., et\u00a0al.: Sentry: Selective entropy optimization via committee consistency for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8558\u20138567 (2021)","DOI":"10.1109\/ICCV48922.2021.00844"},{"key":"1444_CR47","doi-asserted-by":"crossref","unstructured":"Qiu, Z., Zhang, Y., Lin, H., et\u00a0al.: Source-free domain adaptation via avatar prototype generation and adaptation. In: International Joint Conference on Artificial Intelligence, pp. 2921\u20132927 (2021)","DOI":"10.24963\/ijcai.2021\/402"},{"key":"1444_CR48","doi-asserted-by":"crossref","unstructured":"Saenko, K., Kulis, B., Fritz, M., et\u00a0al.: Adapting visual category models to new domains. In: European Conference on Computer Vision, Springer, pp. 213\u2013226 (2010)","DOI":"10.1007\/978-3-642-15561-1_16"},{"key":"1444_CR49","doi-asserted-by":"crossref","unstructured":"Saito, K., Watanabe, K., et\u00a0al.: Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3723\u20133732 (2018)","DOI":"10.1109\/CVPR.2018.00392"},{"issue":"1","key":"1444_CR50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-019-0197-0","volume":"6","author":"C Shorten","year":"2019","unstructured":"Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. J. Big Data 6(1), 1\u201348 (2019). https:\/\/doi.org\/10.1186\/s40537-019-0197-0","journal-title":"J. Big Data"},{"key":"1444_CR51","unstructured":"Sohn, K., Berthelot, D., Carlini, N., et\u00a0al.: Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In: Neural Information Processing Systems, pp. 596\u2013608 (2020)"},{"key":"1444_CR52","doi-asserted-by":"crossref","unstructured":"Sun, B., Feng, J., Saenko, K.: Return of frustratingly easy domain adaptation. In: AAAI Conference on Artificial Intelligence, pp. 2058\u20132065 (2016)","DOI":"10.1609\/aaai.v30i1.10306"},{"key":"1444_CR53","doi-asserted-by":"crossref","unstructured":"Tang, H., Chen, K., Jia, K.: Unsupervised domain adaptation via structurally regularized deep clustering. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8722\u20138732 (2020)","DOI":"10.1109\/CVPR42600.2020.00875"},{"key":"1444_CR54","doi-asserted-by":"crossref","unstructured":"Tang, S., Shi, Y., Ma, Z., et\u00a0al.: Model adaptation through hypothesis transfer with gradual knowledge distillation. In: 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, pp. 5679\u20135685 (2021)","DOI":"10.1109\/IROS51168.2021.9636206"},{"key":"1444_CR55","doi-asserted-by":"publisher","first-page":"467","DOI":"10.1016\/j.neunet.2022.05.015","volume":"152","author":"S Tang","year":"2022","unstructured":"Tang, S., Zou, Y., Song, Z., et al.: Semantic consistency learning on manifold for source data-free unsupervised domain adaptation. Neural Netw. 152, 467\u2013478 (2022). https:\/\/doi.org\/10.1016\/j.neunet.2022.05.015","journal-title":"Neural Netw."},{"issue":"6","key":"1444_CR56","doi-asserted-by":"publisher","first-page":"3749","DOI":"10.1109\/TCSVT.2021.3111034","volume":"32","author":"J Tian","year":"2021","unstructured":"Tian, J., Zhang, J., Li, W., et al.: Vdm-da: Virtual domain modeling for source data-free domain adaptation. IEEE Trans. Circuits Syst. Video Technol. 32(6), 3749\u20133760 (2021). https:\/\/doi.org\/10.1109\/TCSVT.2021.3111034","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1444_CR57","doi-asserted-by":"crossref","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. In: European Conference on Computer Vision, pp. 776\u2013794 (2020)","DOI":"10.1007\/978-3-030-58621-8_45"},{"key":"1444_CR58","doi-asserted-by":"crossref","unstructured":"Venkateswara, H., Eusebio, J., Chakraborty, S., et\u00a0al.: Deep hashing network for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5018\u20135027 (2017)","DOI":"10.1109\/CVPR.2017.572"},{"key":"1444_CR59","unstructured":"Wang, D., Shelhamer, E., Liu, S., et\u00a0al.: Tent: Fully test-time adaptation by entropy minimization. In: International Conference on Learning Representations (2021)"},{"key":"1444_CR60","doi-asserted-by":"crossref","unstructured":"Wang, F., Han, Z., Gong, Y., et\u00a0al.: Exploring domain-invariant parameters for source free domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7151\u20137160 (2022)","DOI":"10.1109\/CVPR52688.2022.00701"},{"key":"1444_CR61","doi-asserted-by":"crossref","unstructured":"Wu, Z., Xiong, Y., Yu, S.X., et\u00a0al.: Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3733\u20133742 (2018)","DOI":"10.1109\/CVPR.2018.00393"},{"key":"1444_CR62","doi-asserted-by":"crossref","unstructured":"Xie, Q., Luong, M.T., Hovy, E., et\u00a0al.: Self-training with noisy student improves imagenet classification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10687\u201310698 (2020)","DOI":"10.1109\/CVPR42600.2020.01070"},{"key":"1444_CR63","doi-asserted-by":"crossref","unstructured":"Xu, M., Wang, H., Ni, B., et\u00a0al.: Cross-domain detection via graph-induced prototype alignment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12355\u201312364 (2020)","DOI":"10.1109\/CVPR42600.2020.01237"},{"key":"1444_CR64","doi-asserted-by":"crossref","unstructured":"Xu, R., Li, G., Yang, J., et\u00a0al.: Larger norm more transferable: An adaptive feature norm app.roach for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1426\u20131435 (2019)","DOI":"10.1109\/ICCV.2019.00151"},{"key":"1444_CR65","doi-asserted-by":"crossref","unstructured":"Yang, G., Xia, H., et\u00a0al.: Bi-directional generation for unsupervised domain adaptation. In: AAAI Conference on Artificial Intelligence, pp. 6615\u20136622 (2020a)","DOI":"10.1609\/aaai.v34i04.6137"},{"key":"1444_CR66","unstructured":"Yang, S., Wang, Y., Van De\u00a0Weijer, J., et\u00a0al.: Unsupervised domain adaptation without source data by casting a bait. (2020b). arXiv preprint arXiv:2010.12427"},{"key":"1444_CR67","doi-asserted-by":"crossref","unstructured":"Yang, S., Wang, Y., Van De\u00a0Weijer, J., et\u00a0al.: Generalized source-free domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8978\u20138987 (2021a)","DOI":"10.1109\/ICCV48922.2021.00885"},{"key":"1444_CR68","unstructured":"Yang, S., van\u00a0de Weijer, J., Herranz, L., et\u00a0al.: Exploiting the intrinsic neighborhood structure for source-free domain adaptation. In: Neural Information Processing Systems, pp. 29393\u201329405 (2021b)"},{"key":"1444_CR69","doi-asserted-by":"crossref","unstructured":"Ye, M., Zhang, X., Yuen, P.C., et\u00a0al.: Unsupervised embedding learning via invariant and spreading instance feature. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6210\u20136219 (2019)","DOI":"10.1109\/CVPR.2019.00637"},{"key":"1444_CR70","unstructured":"Zellinger, W., Grubinger, T., Lughofer, E., et\u00a0al.: Central moment discrepancy (cmd) for domain-invariant representation learning. In: International Conference on Learning Representations (2017)"},{"key":"1444_CR71","unstructured":"Zhang, Y., Liu, T., et\u00a0al.: Bridging theory and algorithm for domain adaptation. In: International Conference on Machine Learning, pp. 7404\u20137413 (2019)"},{"issue":"7","key":"1444_CR72","doi-asserted-by":"publisher","first-page":"7862","DOI":"10.1007\/s10489-022-03805-9","volume":"53","author":"X Zhao","year":"2023","unstructured":"Zhao, X., Wang, S., Sun, Q.: Open-set domain adaptation by deconfounding domain gaps. Appl. Intell. 53(7), 7862\u20137875 (2023). https:\/\/doi.org\/10.1007\/s10489-022-03805-9","journal-title":"Appl. Intell."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01444-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01444-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01444-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T18:09:35Z","timestamp":1730138975000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01444-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,16]]},"references-count":72,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1444"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01444-3","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2024,8,16]]},"assertion":[{"value":"14 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 August 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 August 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"248"}}