{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T16:38:29Z","timestamp":1742920709431,"version":"3.40.3"},"publisher-location":"Cham","reference-count":44,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031587757"},{"type":"electronic","value":"9783031587764"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-58776-4_18","type":"book-chapter","created":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T14:04:33Z","timestamp":1719842673000},"page":"222-235","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["3D Swin Transformer for\u00a0Partial Medical Auto Segmentation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0079-9495","authenticated-orcid":false,"given":"Aneesh","family":"Rangnekar","sequence":"first","affiliation":[]},{"given":"Jue","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Harini","family":"Veeraraghavan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,1]]},"reference":[{"issue":"3","key":"18_CR1","doi-asserted-by":"publisher","first-page":"1686","DOI":"10.1002\/mp.15507","volume":"49","author":"A Amjad","year":"2022","unstructured":"Amjad, A., et al.: General and custom deep learning autosegmentation models for organs in head and neck, abdomen, and male pelvis. Med. Phys. 49(3), 1686\u20131700 (2022)","journal-title":"Med. Phys."},{"issue":"1","key":"18_CR2","doi-asserted-by":"publisher","first-page":"4128","DOI":"10.1038\/s41467-022-30695-9","volume":"13","author":"M Antonelli","year":"2022","unstructured":"Antonelli, M., et al.: The medical segmentation decathlon. Nat. Commun. 13(1), 4128 (2022)","journal-title":"Nat. Commun."},{"key":"18_CR3","unstructured":"Bao, H., Dong, L., Piao, S., Wei, F.: BEiT: BERT pre-training of image transformers. arXiv preprint arXiv:2106.08254 (2021)"},{"key":"18_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102680","volume":"84","author":"P Bilic","year":"2023","unstructured":"Bilic, P., et al.: The liver tumor segmentation benchmark (LiTS). Med. Image Anal. 84, 102680 (2023)","journal-title":"Med. Image Anal."},{"key":"18_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"220","DOI":"10.1007\/978-3-031-19806-9_13","volume-title":"Computer Vision","author":"YH Cao","year":"2022","unstructured":"Cao, Y.H., Yu, H., Wu, J.: Training vision transformers with only 2040 images. In: Avidan, S., Brostow, G., Cisse, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13685, pp. 220\u2013237. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19806-9_13"},{"issue":"6","key":"18_CR6","doi-asserted-by":"publisher","first-page":"1045","DOI":"10.1007\/s10278-013-9622-7","volume":"26","author":"K Clark","year":"2013","unstructured":"Clark, K., et al.: The cancer imaging archive (TCIA): maintaining and operating a public information repository. J. Digit. Imaging 26(6), 1045\u20131057 (2013)","journal-title":"J. Digit. Imaging"},{"key":"18_CR7","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth $$16 \\times 16$$ words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Fan, H., et al.: Multiscale vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6824\u20136835 (2021)","DOI":"10.1109\/ICCV48922.2021.00675"},{"key":"18_CR9","doi-asserted-by":"publisher","unstructured":"Gatidis, S., et\u00a0al.: The autopet challenge: towards fully automated lesion segmentation in oncologic PET\/CT imaging. Preprint at Research Square (Nature Portfolio) (2023). https:\/\/doi.org\/10.21203\/rs.3.rs-2572595\/v1","DOI":"10.21203\/rs.3.rs-2572595\/v1"},{"issue":"1","key":"18_CR10","doi-asserted-by":"publisher","first-page":"601","DOI":"10.1038\/s41597-022-01718-3","volume":"9","author":"S Gatidis","year":"2022","unstructured":"Gatidis, S., et al.: A whole-body FDG-PET\/CT dataset with manually annotated tumor lesions. Sci. Data 9(1), 601 (2022)","journal-title":"Sci. Data"},{"key":"18_CR11","unstructured":"Goyal, P., et al.: Accurate, large minibatch SGD: training imagenet in 1 hour. arXiv preprint arXiv:1706.02677 (2017)"},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"18_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2020.101821","volume":"67","author":"N Heller","year":"2021","unstructured":"Heller, N., et al.: The state of the art in kidney and kidney tumor segmentation in contrast-enhanced CT imaging: results of the KiTS19 challenge. Med. Image Anal. 67, 101821 (2021)","journal-title":"Med. Image Anal."},{"issue":"6","key":"18_CR14","doi-asserted-by":"publisher","first-page":"626","DOI":"10.1200\/JCO.2020.38.6_suppl.626","volume":"38","author":"N Heller","year":"2020","unstructured":"Heller, N., et al.: An international challenge to use artificial intelligence to define the state-of-the-art in kidney and kidney tumor segmentation in CT imaging. Proc. Am. Soc. Clin. Oncol. 38(6), 626 (2020)","journal-title":"Proc. Am. Soc. Clin. Oncol."},{"key":"18_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"178","DOI":"10.1007\/978-3-031-23911-3_16","volume-title":"Fast and Low-Resource Semi-supervised Abdominal Organ Segmentation","author":"Z Huang","year":"2022","unstructured":"Huang, Z., et al.: Revisiting nnU-net for iterative pseudo labeling and efficient sliding window inference. In: Ma, J., Wang, B. (eds.) FLARE 2022. LNCS, vol. 13816, pp. 178\u2013189. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-23911-3_16"},{"issue":"2","key":"18_CR16","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"issue":"8","key":"18_CR17","doi-asserted-by":"publisher","first-page":"5244","DOI":"10.1002\/mp.15765","volume":"49","author":"J Jiang","year":"2022","unstructured":"Jiang, J., et al.: Nested block self-attention multiple resolution residual network for multiorgan segmentation from CT. Med. Phys. 49(8), 5244\u20135257 (2022)","journal-title":"Med. Phys."},{"key":"18_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"556","DOI":"10.1007\/978-3-031-16440-8_53","volume-title":"Medical Image Computing and Computer Assisted Intervention","author":"J Jiang","year":"2022","unstructured":"Jiang, J., Tyagi, N., Tringale, K., Crane, C., Veeraraghavan, H.: Self-supervised 3D anatomy segmentation using self-distilled masked image transformer (SMIT). In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) MICCAI 2022. LNCS, vol. 13434, pp. 556\u2013566. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16440-8_53"},{"key":"18_CR19","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)"},{"key":"18_CR20","unstructured":"Landman, B., Xu, Z., Igelsias, J., Styner, M., Langerak, T., Klein, A.: MICCAI multi-atlas labeling beyond the cranial vault\u2013workshop and challenge. In: Proceedings of the MICCAI Multi-Atlas Labeling Beyond Cranial Vault-Workshop Challenge, vol.\u00a05, p.\u00a012 (2015)"},{"key":"18_CR21","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"18_CR22","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)"},{"key":"18_CR23","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"18_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102035","volume":"71","author":"J Ma","year":"2021","unstructured":"Ma, J., et al.: Loss odyssey in medical image segmentation. Med. Image Anal. 71, 102035 (2021)","journal-title":"Med. Image Anal."},{"key":"18_CR25","doi-asserted-by":"publisher","first-page":"654","DOI":"10.1038\/s41467-024-44824-z","volume":"15","author":"J Ma","year":"2024","unstructured":"Ma, J., He, Y., Li, F., Han, L., You, C., Wang, B.: Segment anything in medical images. Nat. Commun. 15, 654 (2024)","journal-title":"Nat. Commun."},{"key":"18_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102616","volume":"82","author":"J Ma","year":"2022","unstructured":"Ma, J., et al.: Fast and low-GPU-memory abdomen CT organ segmentation: the flare challenge. Med. Image Anal. 82, 102616 (2022)","journal-title":"Med. Image Anal."},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Ma, J., et al.: Unleashing the strengths of unlabeled data in pan-cancer abdominal organ quantification: the flare22 challenge. arXiv preprint arXiv:2308.05862 (2023)","DOI":"10.1016\/S2589-7500(24)00154-7"},{"issue":"10","key":"18_CR28","doi-asserted-by":"publisher","first-page":"6695","DOI":"10.1109\/TPAMI.2021.3100536","volume":"44","author":"J Ma","year":"2022","unstructured":"Ma, J., et al.: AbdomenCT-1K: is abdominal organ segmentation a solved problem? IEEE Trans. Pattern Anal. Mach. Intell. 44(10), 6695\u20136714 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Olsson, V., Tranheden, W., Pinto, J., Svensson, L.: Classmix: segmentation-based data augmentation for semi-supervised learning. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 1369\u20131378 (2021)","DOI":"10.1109\/WACV48630.2021.00141"},{"key":"18_CR30","unstructured":"Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d\u2019 Alch\u00e9-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems 32, pp. 8024\u20138035. Curran Associates, Inc. (2019). http:\/\/papers.neurips.cc\/paper\/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf"},{"issue":"198","key":"18_CR31","first-page":"1","volume":"24","author":"A Pavao","year":"2023","unstructured":"Pavao, A., et al.: CodaLab competitions: an open source platform to organize scientific challenges. J. Mach. Learn. Res. 24(198), 1\u20136 (2023)","journal-title":"J. Mach. Learn. Res."},{"key":"18_CR32","doi-asserted-by":"crossref","unstructured":"Rangnekar, A., Kanan, C., Hoffman, M.: Semantic segmentation with active semi-supervised representation learning. arXiv preprint arXiv:2210.08403 (2022)","DOI":"10.1109\/WACV56688.2023.00591"},{"key":"18_CR33","unstructured":"Simpson, A.L., et al.: A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv preprint arXiv:1902.09063 (2019)"},{"key":"18_CR34","doi-asserted-by":"crossref","unstructured":"Tang, Y., et al.: Self-supervised pre-training of swin transformers for 3D medical image analysis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20730\u201320740 (2022)","DOI":"10.1109\/CVPR52688.2022.02007"},{"key":"18_CR35","unstructured":"Tarvainen, A., Valpola, H.: Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 1195\u20131204. Curran Associates Inc., Red Hook (2017)"},{"key":"18_CR36","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp. 10347\u201310357. PMLR (2021)"},{"key":"18_CR37","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1016\/j.radonc.2020.09.008","volume":"153","author":"L Vandewinckele","year":"2020","unstructured":"Vandewinckele, L., et al.: Overview of artificial intelligence-based applications in radiotherapy: recommendations for implementation and quality assurance. Radiother. Oncol. 153, 55\u201366 (2020)","journal-title":"Radiother. Oncol."},{"key":"18_CR38","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"202","DOI":"10.1007\/978-3-031-23911-3_18","volume-title":"Fast and Low-Resource Semi-supervised Abdominal Organ Segmentation","author":"E Wang","year":"2022","unstructured":"Wang, E., Zhao, Y., Wu, Y.: Cascade dual-decoders network for abdominal organs segmentation. In: Ma, J., Wang, B. (eds.) FLARE 2022. LNCS, vol. 13816, pp. 202\u2013213. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-23911-3_18"},{"key":"18_CR39","doi-asserted-by":"crossref","unstructured":"Wasserthal, J., et al.: TotalSegmentator: robust segmentation of 104 anatomic structures in CT images. Radiol. Artif. Intelli. 5(5), e230024 (2023)","DOI":"10.1148\/ryai.230024"},{"key":"18_CR40","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"605","DOI":"10.1007\/978-3-031-20056-4_35","volume-title":"Computer Vision","author":"Z Weng","year":"2022","unstructured":"Weng, Z., Yang, X., Li, A., Wu, Z., Jiang, Y.G.: Semi-supervised vision transformers. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13690, pp. 605\u2013620. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20056-4_35"},{"key":"18_CR41","doi-asserted-by":"crossref","unstructured":"Xie, Q., Luong, M.T., Hovy, E., Le, Q.V.: Self-training with noisy student improves imagenet classification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10687\u201310698 (2020)","DOI":"10.1109\/CVPR42600.2020.01070"},{"key":"18_CR42","doi-asserted-by":"crossref","unstructured":"Yang, L., Zhuo, W., Qi, L., Shi, Y., Gao, Y.: ST++: make self-training work better for semi-supervised semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.00423"},{"key":"18_CR43","doi-asserted-by":"crossref","unstructured":"Yun, S., Han, D., Oh, S.J., Chun, S., Choe, J., Yoo, Y.: CutMix: regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6023\u20136032 (2019)","DOI":"10.1109\/ICCV.2019.00612"},{"key":"18_CR44","doi-asserted-by":"crossref","unstructured":"Yushkevich, P.A., Gao, Y., Gerig, G.: ITK-snap: an interactive tool for semi-automatic segmentation of multi-modality biomedical images. In: Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 3342\u20133345 (2016)","DOI":"10.1109\/EMBC.2016.7591443"}],"container-title":["Lecture Notes in Computer Science","Fast, Low-resource, and Accurate Organ and Pan-cancer Segmentation in Abdomen CT"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-58776-4_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T05:37:45Z","timestamp":1732340265000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-58776-4_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031587757","9783031587764"],"references-count":44,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-58776-4_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 July 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"FLARE","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"MICCAI Challenge on Fast and Low-Resource Semi-supervised Abdominal Organ Segmentation","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"flare2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/codalab.lisn.upsaclay.fr\/competitions\/12239","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}