{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T12:26:17Z","timestamp":1775737577184,"version":"3.50.1"},"publisher-location":"Cham","reference-count":41,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031917202","type":"print"},{"value":"9783031917219","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-91721-9_4","type":"book-chapter","created":{"date-parts":[[2025,5,29]],"date-time":"2025-05-29T05:03:55Z","timestamp":1748495035000},"page":"51-67","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":56,"title":["SAM-Med3D: Towards General-Purpose Segmentation Models for\u00a0Volumetric Medical Images"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1753-7336","authenticated-orcid":false,"given":"Haoyu","family":"Wang","sequence":"first","affiliation":[]},{"given":"Sizheng","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Zhongying","family":"Deng","sequence":"additional","affiliation":[]},{"given":"Junlong","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Tianbin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jianpin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yanzhou","family":"Su","sequence":"additional","affiliation":[]},{"given":"Ziyan","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Yiqing","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Shaoting","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1813-1784","authenticated-orcid":false,"given":"Junjun","family":"He","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1889-2567","authenticated-orcid":false,"given":"Yu","family":"Qiao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"key":"4_CR1","unstructured":"Baid, U., et al.: The RSNA-ASNR-MICCAI brats 2021 benchmark on brain tumor segmentation and radiogenomic classification. CoRR abs\/2107.02314 (2021). https:\/\/arxiv.org\/abs\/2107.02314"},{"key":"4_CR2","doi-asserted-by":"crossref","unstructured":"Bui, N.T., Hoang, D.H., Tran, M.T., Le, N.: SAM3D: segment anything model in volumetric medical images. arXiv preprint arXiv:2309.03493 (2023)","DOI":"10.1109\/ISBI56570.2024.10635844"},{"key":"4_CR3","doi-asserted-by":"crossref","unstructured":"Chen, C., et\u00a0al.: MA-SAM: modality-agnostic SAM adaptation for 3D medical image segmentation. arXiv preprint arXiv:2309.08842 (2023)","DOI":"10.1016\/j.media.2024.103310"},{"key":"4_CR4","unstructured":"Chen, H., et al.: Cerebral artery segmentation challenge. In: International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) (2023)"},{"key":"4_CR5","unstructured":"Cheng, D., Qin, Z., Jiang, Z., Zhang, S., Lao, Q., Li, K.: SAM on medical images: a comprehensive study on three prompt modes (2023)"},{"key":"4_CR6","unstructured":"Cheng, J., et\u00a0al.: SAM-Med2D. arXiv preprint arXiv:2308.16184 (2023)"},{"key":"4_CR7","unstructured":"Deng, R., et al.: Segment anything model (SAM) for digital pathology: assess zero-shot segmentation on whole slide imaging (2023)"},{"key":"4_CR8","unstructured":"Du, Y., Bai, F., Huang, T., Zhao, B.: SegVol: universal and interactive volumetric medical image segmentation. arXiv preprint arXiv:2311.13385 (2023)"},{"key":"4_CR9","doi-asserted-by":"crossref","unstructured":"Gatidis, S., et\u00a0al.: The autopet challenge: towards fully automated lesion segmentation in oncologic PET\/CT imaging (2023)","DOI":"10.21203\/rs.3.rs-2572595\/v1"},{"key":"4_CR10","doi-asserted-by":"crossref","unstructured":"Gong, S., et al.: 3DSAM-adapter: holistic adaptation of SAM from 2D to 3D for promptable medical image segmentation. arXiv preprint arXiv:2306.13465 (2023)","DOI":"10.1016\/j.media.2024.103324"},{"key":"4_CR11","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A., et al.: UNETR: transformers for 3d medical image segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 574\u2013584 (2022)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"4_CR12","unstructured":"Heller, N., et al.: The KITS21 challenge: automatic segmentation of kidneys, renal tumors, and renal cysts in corticomedullary-phase CT (2023)"},{"key":"4_CR13","unstructured":"Hu, C., Xia, T., Ju, S., Li, X.: When SAM meets medical images: an investigation of segment anything model (SAM) on multi-phase liver tumor segmentation (2023)"},{"key":"4_CR14","unstructured":"Huang, Y., et al.: Segment anything model for medical images? (2023)"},{"key":"4_CR15","unstructured":"Huang, Z., et\u00a0al.: STU-Net: scalable and transferable medical image segmentation models empowered by large-scale supervised pre-training. arXiv preprint arXiv:2304.06716 (2023)"},{"issue":"2","key":"4_CR16","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"4_CR17","unstructured":"Ji, Y., et\u00a0al.: AMOS: a large-scale abdominal multi-organ benchmark for versatile medical image segmentation. arXiv preprint arXiv:2206.08023 (2022)"},{"key":"4_CR18","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)"},{"key":"4_CR19","unstructured":"Landman, B., Xu, Z., Igelsias, J., Styner, M., Langerak, T., Klein, A.: MICCAI multi-atlas labeling beyond the cranial vault\u2013workshop and challenge. In: Proceedings of the MICCAI Multi-Atlas Labeling Beyond Cranial Vault-Workshop Challenge, vol.\u00a05, p.\u00a012 (2015)"},{"key":"4_CR20","unstructured":"Lei, W., Wei, X., Zhang, X., Li, K., Zhang, S.: MedLSAM: localize and segment anything model for 3d medical images. arXiv preprint arXiv:2306.14752 (2023)"},{"issue":"12","key":"4_CR21","doi-asserted-by":"publisher","first-page":"2663","DOI":"10.1109\/TMI.2018.2845918","volume":"37","author":"X Li","year":"2018","unstructured":"Li, X., Chen, H., Qi, X., Dou, Q., Fu, C.W., Heng, P.A.: H-DenseUNet: hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 37(12), 2663\u20132674 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4_CR22","doi-asserted-by":"crossref","unstructured":"Liu, J., et al.: Clip-driven universal model for organ segmentation and tumor detection. arXiv preprint arXiv:2301.00785 (2023)","DOI":"10.1109\/ICCV51070.2023.01934"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"L\u00f6ffler, M.T., et al.: A vertebral segmentation dataset with fracture grading. Radiol. Artif. Intell. 2(4), e190138 (2020)","DOI":"10.1148\/ryai.2020190138"},{"key":"4_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102102","volume":"72","author":"X Luo","year":"2021","unstructured":"Luo, X., et al.: MIDeepSeg: minimally interactive segmentation of unseen objects from medical images using deep learning. Med. Image Anal. 72, 102102 (2021)","journal-title":"Med. Image Anal."},{"key":"4_CR25","doi-asserted-by":"crossref","unstructured":"Ma, J., Wang, B.: Segment anything in medical images. arXiv preprint arXiv:2304.12306 (2023)","DOI":"10.1038\/s41467-024-44824-z"},{"key":"4_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102918","volume":"89","author":"MA Mazurowski","year":"2023","unstructured":"Mazurowski, M.A., Dong, H., Gu, H., Yang, J., Konz, N., Zhang, Y.: Segment anything model for medical image analysis: an experimental study. Med. Image Anal. 89, 102918 (2023)","journal-title":"Med. Image Anal."},{"issue":"1","key":"4_CR27","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1038\/s41597-021-00946-3","volume":"8","author":"K Payette","year":"2021","unstructured":"Payette, K., et al.: An automatic multi-tissue human fetal brain segmentation benchmark using the fetal tissue annotation dataset. Sci. Data 8(1), 167 (2021)","journal-title":"Sci. Data"},{"key":"4_CR28","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision (2021)"},{"key":"4_CR29","doi-asserted-by":"crossref","unstructured":"Radl, L., et al.: AVT: multicenter aortic vessel tree CTA dataset collection with ground truth segmentation masks. Data Brief 40, 107801 (2022)","DOI":"10.1016\/j.dib.2022.107801"},{"key":"4_CR30","unstructured":"Ramesh, A., et al.: Zero-shot text-to-image generation (2021)"},{"key":"4_CR31","unstructured":"Wang, C., et al.: SAMMed: a medical image annotation framework based on large vision model. arXiv preprint arXiv:2307.05617 (2023)"},{"key":"4_CR32","doi-asserted-by":"crossref","unstructured":"Wang, H., Yi, H., Liu, J., Gu, L.: Integrated treatment planning in percutaneous microwave ablation of lung tumors. In: 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pp. 4974\u20134977. IEEE (2022)","DOI":"10.1109\/EMBC48229.2022.9871915"},{"key":"4_CR33","unstructured":"Wang, K.: Tumor detection, segmentation and classification challenge on automated 3D breast ultrasound (ABUS) 2023 (2023). https:\/\/tdsc-abus2023.grand-challenge.org\/TDSC-ABUS2023\/"},{"issue":"9","key":"4_CR34","doi-asserted-by":"publisher","first-page":"2463","DOI":"10.1109\/TMI.2021.3079709","volume":"40","author":"X Wang","year":"2021","unstructured":"Wang, X., et al.: Joint learning of 3D lesion segmentation and classification for explainable COVID-19 diagnosis. IEEE Trans. Med. Imaging 40(9), 2463\u20132476 (2021). https:\/\/doi.org\/10.1109\/TMI.2021.3079709","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4_CR35","doi-asserted-by":"crossref","unstructured":"Wang, X., Zhang, X., Cao, Y., Wang, W., Shen, C., Huang, T.: SegGPT: segmenting everything in context. arXiv preprint arXiv:2304.03284 (2023)","DOI":"10.1109\/ICCV51070.2023.00110"},{"key":"4_CR36","doi-asserted-by":"crossref","unstructured":"Wasserthal, J., et al.: TotalSegmentator: robust segmentation of 104 anatomic structures in CT images. Radiol. Artif. Intelli. 5(5), e230024 (2023)","DOI":"10.1148\/ryai.230024"},{"key":"4_CR37","unstructured":"Wu, J., et al.: Medical SAM adapter: adapting segment anything model for medical image segmentation. arXiv preprint arXiv:2304.12620 (2023)"},{"key":"4_CR38","unstructured":"Yang, T., Zhu, Y., Xie, Y., Zhang, A., Chen, C., Li, M.: AIM: adapting image models for efficient video action recognition. arXiv preprint arXiv:2302.03024 (2023)"},{"key":"4_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, J., Xie, Y., Xia, Y., Shen, C.: DoDNet: learning to segment multi-organ and tumors from multiple partially labeled datasets. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1195\u20131204 (2021)","DOI":"10.1109\/CVPR46437.2021.00125"},{"key":"4_CR40","unstructured":"Zhou, T., Zhang, Y., Zhou, Y., Wu, Y., Gong, C.: Can SAM segment polyps? (2023)"},{"key":"4_CR41","unstructured":"Zou, X., et al.: Segment everything everywhere all at once (2023)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-91721-9_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,29]],"date-time":"2025-05-29T05:04:16Z","timestamp":1748495056000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-91721-9_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031917202","9783031917219"],"references-count":41,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-91721-9_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}