{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:40:11Z","timestamp":1742913611549,"version":"3.40.3"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031456725"},{"type":"electronic","value":"9783031456732"}],"license":[{"start":{"date-parts":[[2023,10,15]],"date-time":"2023-10-15T00:00:00Z","timestamp":1697328000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,10,15]],"date-time":"2023-10-15T00:00:00Z","timestamp":1697328000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-45673-2_30","type":"book-chapter","created":{"date-parts":[[2023,10,14]],"date-time":"2023-10-14T08:02:16Z","timestamp":1697270536000},"page":"299-309","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Self-training with\u00a0Domain-Mixed Data for\u00a0Few-Shot Domain Adaptation in\u00a0Medical Image Segmentation Tasks"],"prefix":"10.1007","author":[{"given":"Yongze","family":"Wang","sequence":"first","affiliation":[]},{"given":"Maurice","family":"Pagnucco","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,15]]},"reference":[{"key":"30_CR1","doi-asserted-by":"crossref","unstructured":"Bernal, J., et al.: Comparative validation of polyp detection methods in video colonoscopy: results from the MICCAI 2015 endoscopic vision challenge. IEEE Trans. Med. Imaging 36(6), 1231\u20131249 (2017)","DOI":"10.1109\/TMI.2017.2664042"},{"key":"30_CR2","doi-asserted-by":"publisher","unstructured":"Chen, C., Liu, Q., Jin, Y., Dou, Q., Heng, P.-A.: Source-free domain adaptive fundus image segmentation with denoised pseudo-labeling. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12905, pp. 225\u2013235. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87240-3_22","DOI":"10.1007\/978-3-030-87240-3_22"},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Chen, L.C., et al.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: ECCV, pp. 801\u2013818 (2018)","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"30_CR4","doi-asserted-by":"publisher","unstructured":"Cho, H., Nishimura, K., Watanabe, K., Bise, R.: Cell detection in domain shift problem using pseudo-cell-position heatmap. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12908, pp. 384\u2013394. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87237-3_37","DOI":"10.1007\/978-3-030-87237-3_37"},{"key":"30_CR5","unstructured":"Cordonnier, J.B., Loukas, A., Jaggi, M.: On the relationship between self-attention and convolutional layers. arXiv preprint arXiv:1911.03584 (2019)"},{"key":"30_CR6","doi-asserted-by":"publisher","unstructured":"Dong, N., Kampffmeyer, M., Liang, X., Wang, Z., Dai, W., Xing, E.: Unsupervised domain adaptation for automatic estimation of cardiothoracic ratio. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-L\u00f3pez, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11071, pp. 544\u2013552. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00934-2_61","DOI":"10.1007\/978-3-030-00934-2_61"},{"key":"30_CR7","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"30_CR8","doi-asserted-by":"publisher","unstructured":"Fan, D.-P., et al.: PraNet: parallel reverse attention network for polyp segmentation. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12266, pp. 263\u2013273. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59725-2_26","DOI":"10.1007\/978-3-030-59725-2_26"},{"key":"30_CR9","doi-asserted-by":"publisher","unstructured":"Feng, W., et al.: Unsupervised domain adaptive fundus image segmentation with category-level regularization. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022: 25th International Conference, Singapore, September 18\u201322, 2022, Proceedings, Part II, pp. 497\u2013506. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16434-7_48","DOI":"10.1007\/978-3-031-16434-7_48"},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Fumero, F., Alay\u00f3n, S., Sanchez, J.L.: Rim-one: an open retinal image database for optic nerve evaluation. In: CBMS, pp. 1\u20136. IEEE (2011)","DOI":"10.1109\/CBMS.2011.5999143"},{"key":"30_CR11","unstructured":"Ganin, Y., Lempitsky, V.: Unsupervised domain adaptation by backpropagation. In: International Conference on Machine Learning, pp. 1180\u20131189. PMLR (2015)"},{"key":"30_CR12","unstructured":"Geirhos, R., et al.: Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv preprint arXiv:1811.12231 (2018)"},{"key":"30_CR13","unstructured":"Haq, M.M., Huang, J.: Adversarial domain adaptation for cell segmentation. In: Medical Imaging with Deep Learning, pp. 277\u2013287. PMLR (2020)"},{"key":"30_CR14","doi-asserted-by":"crossref","unstructured":"Hoyer, L., Dai, D., Van Gool, L.: Daformer: improving network architectures and training strategies for domain-adaptive semantic segmentation. In: CVPR, pp. 9924\u20139935 (2022)","DOI":"10.1109\/CVPR52688.2022.00969"},{"key":"30_CR15","doi-asserted-by":"publisher","unstructured":"Hu, M., et al.: Fully test-time adaptation for image segmentation. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12903, pp. 251\u2013260. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87199-4_24","DOI":"10.1007\/978-3-030-87199-4_24"},{"key":"30_CR16","unstructured":"Lee, J., et al.: Confidence score for source-free unsupervised domain adaptation. In: ICML, pp. 12365\u201312377. PMLR (2022)"},{"key":"30_CR17","doi-asserted-by":"publisher","unstructured":"Li, K., Wang, S., Yu, L., Heng, P.-A.: Dual-teacher: integrating intra-domain and inter-domain teachers for annotation-efficient cardiac segmentation. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12261, pp. 418\u2013427. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59710-8_41","DOI":"10.1007\/978-3-030-59710-8_41"},{"key":"30_CR18","doi-asserted-by":"publisher","unstructured":"Li, S., et al.: Few-shot domain adaptation with\u00a0polymorphic transformers. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12902, pp. 330\u2013340. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87196-3_31","DOI":"10.1007\/978-3-030-87196-3_31"},{"key":"30_CR19","unstructured":"Liang, J., Hu, D., Feng, J.: Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In: ICML, pp. 6028\u20136039. PMLR (2020)"},{"key":"30_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: ICCV, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"30_CR21","doi-asserted-by":"crossref","unstructured":"Olsson, V., et al.: Classmix: segmentation-based data augmentation for semi-supervised learning. In: WACV, pp. 1369\u20131378 (2021)","DOI":"10.1109\/WACV48630.2021.00141"},{"key":"30_CR22","doi-asserted-by":"crossref","unstructured":"Orlando, J.I., et al.: Refuge challenge: a unified framework for evaluating automated methods for glaucoma assessment from fundus photographs. Med. Image Anal. 59, 101570 (2020)","DOI":"10.1016\/j.media.2019.101570"},{"key":"30_CR23","doi-asserted-by":"crossref","unstructured":"Pogorelov, K., et al.: Kvasir: a multi-class image dataset for computer aided gastrointestinal disease detection. In: Proceedings of the 8th ACM on Multimedia Systems Conference, pp. 164\u2013169 (2017)","DOI":"10.1145\/3083187.3083212"},{"key":"30_CR24","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"30_CR25","unstructured":"Tarvainen, A., Valpola, H.: Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results. In: NeurIPS 30 (2017)"},{"key":"30_CR26","doi-asserted-by":"crossref","unstructured":"Tranheden, W., et al.: Dacs: domain adaptation via cross-domain mixed sampling. In: WACV, pp. 1379\u20131389 (2021)","DOI":"10.1109\/WACV48630.2021.00142"},{"key":"30_CR27","doi-asserted-by":"crossref","unstructured":"Tzeng, E., Hoffman, J., Saenko, K., Darrell, T.: Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7167\u20137176 (2017)","DOI":"10.1109\/CVPR.2017.316"},{"key":"30_CR28","doi-asserted-by":"publisher","unstructured":"Varsavsky, T., et al.: Test-time unsupervised domain adaptation. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12261, pp. 428\u2013436. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59710-8_42","DOI":"10.1007\/978-3-030-59710-8_42"},{"key":"30_CR29","doi-asserted-by":"crossref","unstructured":"Vu, T.H., Jain, H., Bucher, M., Cord, M., P\u00e9rez, P.: Advent: adversarial entropy minimization for domain adaptation in semantic segmentation. In: CVPR, pp. 2517\u20132526 (2019)","DOI":"10.1109\/CVPR.2019.00262"},{"key":"30_CR30","doi-asserted-by":"publisher","unstructured":"Wang, S., Yu, L., Li, K., Yang, X., Fu, C.-W., Heng, P.-A.: Boundary and entropy-driven adversarial learning for fundus image segmentation. In: Shen, D., et al. (eds.) MICCAI 2019. LNCS, vol. 11764, pp. 102\u2013110. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32239-7_12","DOI":"10.1007\/978-3-030-32239-7_12"},{"issue":"11","key":"30_CR31","doi-asserted-by":"publisher","first-page":"2485","DOI":"10.1109\/TMI.2019.2899910","volume":"38","author":"S Wang","year":"2019","unstructured":"Wang, S., Yu, L., Yang, X., Fu, C.W., Heng, P.A.: Patch-based output space adversarial learning for joint optic disc and cup segmentation. IEEE Trans. Med. Imaging 38(11), 2485\u20132495 (2019)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"30_CR32","doi-asserted-by":"publisher","unstructured":"Wu, S., Chen, C., Xiong, Z., Chen, X., Sun, X.: Uncertainty-aware label rectification for domain adaptive mitochondria segmentation. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12903, pp. 191\u2013200. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87199-4_18","DOI":"10.1007\/978-3-030-87199-4_18"},{"key":"30_CR33","unstructured":"Xie, E., et al.: Segformer: simple and efficient design for semantic segmentation with transformers. NeurIPS 34, 12077\u201312090 (2021)"},{"key":"30_CR34","doi-asserted-by":"publisher","unstructured":"Xu, Z., et al.: Denoising for relaxing: unsupervised domain adaptive fundus image segmentation without source data. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022: 25th International Conference, Singapore, September 18\u201322, 2022, Proceedings, Part V, pp. 214\u2013224. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16443-9_21","DOI":"10.1007\/978-3-031-16443-9_21"},{"key":"30_CR35","doi-asserted-by":"crossref","unstructured":"Zhang, P., Zhang, B., Zhang, T., Chen, D., Wang, Y., Wen, F.: Prototypical pseudo label denoising and target structure learning for domain adaptive semantic segmentation. In: CVPR, pp. 12414\u201312424 (2021)","DOI":"10.1109\/CVPR46437.2021.01223"},{"key":"30_CR36","doi-asserted-by":"publisher","unstructured":"Zhao, Z., Xu, K., Li, S., Zeng, Z., Guan, C.: MT-UDA: towards unsupervised cross-modality medical image segmentation with limited source labels. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 293\u2013303. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_28","DOI":"10.1007\/978-3-030-87193-2_28"},{"key":"30_CR37","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: ICCV, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Machine Learning in Medical Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-45673-2_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T16:57:20Z","timestamp":1710349040000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-45673-2_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,15]]},"ISBN":["9783031456725","9783031456732"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-45673-2_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,10,15]]},"assertion":[{"value":"15 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MLMI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Machine Learning in Medical Imaging","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mlmi-med2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/sites.google.com\/view\/mlmi2023?pli=1","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"139","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"93","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"67% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}