{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T02:10:12Z","timestamp":1769566212648,"version":"3.49.0"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,8]],"date-time":"2026-01-08T00:00:00Z","timestamp":1767830400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T00:00:00Z","timestamp":1769472000000},"content-version":"vor","delay-in-days":19,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100002957","name":"Technische Universit\u00e4t Dresden","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100002957","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["BMC Med Imaging"],"DOI":"10.1186\/s12880-025-02075-4","type":"journal-article","created":{"date-parts":[[2026,1,8]],"date-time":"2026-01-08T14:29:46Z","timestamp":1767882586000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Reducing manual workload in CT and MRI annotation with the Segment Anything Model 2"],"prefix":"10.1186","volume":"26","author":[{"given":"Leo","family":"Misera","sequence":"first","affiliation":[]},{"given":"Sven","family":"Nebelung","sequence":"additional","affiliation":[]},{"given":"Zunamys I.","family":"Carrero","sequence":"additional","affiliation":[]},{"given":"Keno","family":"Bressem","sequence":"additional","affiliation":[]},{"given":"Marta","family":"Ligero","sequence":"additional","affiliation":[]},{"given":"Jens-Peter","family":"K\u00fchn","sequence":"additional","affiliation":[]},{"given":"Ralf-Thorsten","family":"Hoffmann","sequence":"additional","affiliation":[]},{"given":"Daniel","family":"Truhn","sequence":"additional","affiliation":[]},{"given":"Jakob Nikolas","family":"Kather","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,8]]},"reference":[{"key":"2075_CR1","doi-asserted-by":"publisher","first-page":"849","DOI":"10.3390\/jcm13030849","volume":"13","author":"D Aiudi","year":"2024","unstructured":"Aiudi D, Iacoangeli A, Dobran M, Polonara G, Chiapponi M, Mattioli A, et al. The prognostic role of volumetric MRI evaluation in the surgical treatment of glioblastoma. J Clin Med. 2024;13:849.","journal-title":"J Clin Med"},{"key":"2075_CR2","doi-asserted-by":"publisher","first-page":"1795","DOI":"10.1007\/s00330-020-07147-3","volume":"31","author":"S Koitka","year":"2021","unstructured":"Koitka S, Kroll L, Malamutmann E, Oezcelik A, Nensa F. Fully automated body composition analysis in routine CT imaging using 3D semantic segmentation convolutional neural networks. Eur Radiol. 2021;31:1795\u2013804.","journal-title":"Eur Radiol"},{"key":"2075_CR3","doi-asserted-by":"publisher","first-page":"4006","DOI":"10.1038\/ncomms5006","volume":"5","author":"HJWL Aerts","year":"2014","unstructured":"Aerts HJWL, Velazquez ER, Leijenaar RTH, Parmar C, Grossmann P, Carvalho S, et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat Commun. 2014;5:4006.","journal-title":"Nat Commun"},{"key":"2075_CR4","doi-asserted-by":"publisher","first-page":"833","DOI":"10.1148\/radiol.2018172300","volume":"287","author":"N Horvat","year":"2018","unstructured":"Horvat N, Veeraraghavan H, Khan M, Blazic I, Zheng J, Capanu M, et al. MR imaging of rectal cancer: radiomics analysis to assess treatment response after neoadjuvant therapy. Radiology. 2018;287:833\u201343.","journal-title":"Radiology"},{"key":"2075_CR5","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.clon.2021.12.003","volume":"34","author":"K Harrison","year":"2022","unstructured":"Harrison K, Pullen H, Welsh C, Oktay O, Alvarez-Valle J, Jena R. Machine learning for auto-segmentation in radiotherapy planning. Clin Oncol (R Coll Radiol). 2022;34:74\u201388.","journal-title":"Clin Oncol (R Coll Radiol)"},{"key":"2075_CR6","doi-asserted-by":"publisher","first-page":"e233029","DOI":"10.1148\/radiol.233029","volume":"314","author":"M Kashyap","year":"2025","unstructured":"Kashyap M, Wang X, Panjwani N, Hasan M, Zhang Q, Huang C, et al. Automated deep learning-based detection and segmentation of lung tumors at ct. Radiology. 2025;314:e233029.","journal-title":"Radiology"},{"key":"2075_CR7","doi-asserted-by":"publisher","first-page":"1529","DOI":"10.1007\/s10994-022-06253-1","volume":"113","author":"Z Pullar-Strecker","year":"2024","unstructured":"Pullar-Strecker Z, Dost K, Frank E, Wicker J. Hitting the target: stopping active learning at the cost-based optimum. Mach Learn. 2024;113:1529\u201347.","journal-title":"Mach Learn"},{"key":"2075_CR8","doi-asserted-by":"publisher","first-page":"950","DOI":"10.1016\/j.ijrobp.2009.09.062","volume":"77","author":"PM Harari","year":"2010","unstructured":"Harari PM, Song S, Tom\u00e9 WA. Emphasizing conformal avoidance versus target definition for IMRT planning in head-and-neck cancer. Int J Radiat Oncol Biol Phys. 2010;77:950\u201358.","journal-title":"Int J Radiat Oncol Biol Phys"},{"key":"2075_CR9","doi-asserted-by":"publisher","first-page":"92","DOI":"10.1016\/j.radonc.2012.02.010","volume":"103","author":"TS Hong","year":"2012","unstructured":"Hong TS, Tom\u00e9 WA, Harari PM. Heterogeneity in head and neck IMRT target design and clinical practice. Radiother Oncol. 2012;103:92\u201398.","journal-title":"Radiother Oncol"},{"key":"2075_CR10","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.radonc.2007.11.016","volume":"87","author":"Z Chen","year":"2008","unstructured":"Chen Z, King W, Pearcey R, Kerba M, Mackillop WJ. The relationship between waiting time for radiotherapy and clinical outcomes: a systematic review of the literature. Radiother Oncol. 2008;87:3\u201316.","journal-title":"Radiother Oncol"},{"key":"2075_CR11","doi-asserted-by":"publisher","first-page":"4128","DOI":"10.1038\/s41467-022-30695-9","volume":"13","author":"M Antonelli","year":"2022","unstructured":"Antonelli M, Reinke A, Bakas S, Farahani K, Kopp-Schneider A, Landman BA, et al. The medical segmentation decathlon. Nat Commun. 2022;13:4128.","journal-title":"Nat Commun"},{"key":"2075_CR12","doi-asserted-by":"crossref","unstructured":"Sun C, Shrivastava A, Singh S, Gupta A. Revisiting unreasonable effectiveness of data in deep learning era. 2017 IEEE International Conference on Computer Vision (ICCV). 2017. p. 843\u201352.","DOI":"10.1109\/ICCV.2017.97"},{"key":"2075_CR13","doi-asserted-by":"crossref","unstructured":"Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, et al. Segment anything. arXiv [cs.CV]. 2023.","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2075_CR14","doi-asserted-by":"publisher","first-page":"e232085","DOI":"10.1148\/radiol.232085","volume":"312","author":"L Misera","year":"2024","unstructured":"Misera L, M\u00fcller-Franzes G, Truhn D, Kather JN. Weakly supervised deep learning in radiology. Radiology. 2024;312:e232085.","journal-title":"Radiology"},{"key":"2075_CR15","doi-asserted-by":"publisher","DOI":"10.1007\/s10278-025-01408-7","author":"Y Zhang","year":"2025","unstructured":"Zhang Y, Zhao S, Gu H, Mazurowski MA. How to efficiently annotate images for best-performing deep learning-based segmentation models: an empirical study with weak and noisy annotations and segment anything model. J Imag Inf Med. 2025. https:\/\/doi.org\/10.1007\/s10278-025-01408-7.","journal-title":"J Imag Inf Med"},{"key":"2075_CR16","unstructured":"Ravi N, Gabeur V, Hu Y-T, Hu R, Ryali C, Ma T, et al. Sam 2: segment anything in images and videos. arXiv [cs.CV]. 2024."},{"key":"2075_CR17","doi-asserted-by":"crossref","unstructured":"Simpson AL, Peoples J, Creasy JM, Fichtinger G, Gangai N, Keshavamurthy KN, et al. Preoperative ct and survival data for patients undergoing resection of colorectal liver metastases. Sci Data. 2024;11:172.","DOI":"10.1038\/s41597-024-02981-2"},{"key":"2075_CR18","doi-asserted-by":"crossref","unstructured":"Simpson AL, Peoples J, Creasy JM, Fichtinger G, Gangai N, Lasso A, et al. Preoperative ct and survival data for patients undergoing resection of colorectal liver metastases (colorectal-liver-metastases). 2023.","DOI":"10.1038\/s41597-024-02981-2"},{"key":"2075_CR19","unstructured":"Dong H, Gu H, Chen Y, Yang J, Chen Y, Mazurowski MA. Segment anything model 2: an application to 2D and 3D medical images. arXiv [cs.CV]. 2024."},{"key":"2075_CR20","doi-asserted-by":"crossref","unstructured":"Ma J, Kim S, Li F, Baharoon M, Asakereh R, Lyu H, et al. Segment anything in medical images and videos: benchmark and deployment. arXiv [eess.IV]. 2024.","DOI":"10.1038\/s41467-024-44824-z"},{"key":"2075_CR21","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.media.2018.11.009","volume":"52","author":"F Ambellan","year":"2019","unstructured":"Ambellan F, Tack A, Ehlke M, Zachow S. Automated segmentation of knee bone and cartilage combining statistical shape knowledge and convolutional neural networks: data from the osteoarthritis initiative. Med Image Anal. 2019;52:109\u201318.","journal-title":"Med Image Anal"},{"key":"2075_CR22","unstructured":"NDA. https:\/\/nda.nih.gov\/oai. Accessed 17 Oct 2024."},{"key":"2075_CR43","doi-asserted-by":"crossref","unstructured":"Yao Y, Chen W. Quantifying knee cartilage shape and lesion: From image to metrics. In: Lecture Notes in Computer Science. Cham: Springer Nature Switzerland; 2025. p. 162\u201372.","DOI":"10.1007\/978-3-031-82007-6_16"},{"key":"2075_CR44","doi-asserted-by":"crossref","unstructured":"Yao Y, Zhong J, Zhang L, Khan S, Chen W. CartiMorph: A framework for automated knee articular cartilage morphometrics. Med Image Anal. 2024;91:103035.","DOI":"10.1016\/j.media.2023.103035"},{"key":"2075_CR23","doi-asserted-by":"publisher","first-page":"102680","DOI":"10.1016\/j.media.2022.102680","volume":"84","author":"P Bilic","year":"2023","unstructured":"Bilic P, Christ P, Li HB, Vorontsov E, Ben-Cohen A, Kaissis G, et al. The liver tumor segmentation benchmark (LiTS). Med Image Anal. 2023;84:102680.","journal-title":"Med Image Anal"},{"key":"2075_CR24","doi-asserted-by":"publisher","first-page":"e240296","DOI":"10.1148\/ryai.240296","volume":"6","author":"K Bartnik","year":"2024","unstructured":"Bartnik K, Bartczak T, Krzyzi\u0144ski M, Korzeniowski K, Lamparski K, W\u0119grzyn P, et al. WAW-TACE: a hepatocellular carcinoma multiphase ct dataset with segmentations, radiomics features, and clinical data. Radiol Artif Intell. 2024;6:e240296.","journal-title":"Radiol Artif Intell"},{"key":"2075_CR25","doi-asserted-by":"publisher","first-page":"108183","DOI":"10.1016\/j.isci.2023.108183","volume":"26","author":"L Wu","year":"2023","unstructured":"Wu L, Wang H, Chen Y, Zhang X, Zhang T, Shen N, et al. Beyond radiologist-level liver lesion detection on multi-phase contrast-enhanced CT images by deep learning. iScience. 2023;26:108183.","journal-title":"iScience"},{"key":"2075_CR26","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee F, Jaeger PF, Kohl SAA, Petersen J, Maier-Hein KH. nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods. 2021;18:203\u201311.","journal-title":"Nat Methods"},{"key":"2075_CR27","doi-asserted-by":"crossref","unstructured":"Wasserthal J, Breit H-C, Meyer MT, Pradella M, Hinck D, Sauter AW, et al. TotalSegmentator: robust segmentation of 104 anatomic structures in CT images. Radiol Artif Intell. 2023;5:e230024.","DOI":"10.1148\/ryai.230024"},{"key":"2075_CR28","doi-asserted-by":"crossref","unstructured":"Maier-Hein L, Reinke A, Godau P, Tizabi MD, Buettner F, Christodoulou E, et al. Metrics reloaded: recommendations for image analysis validation. Nat Methods. 2024;21:195\u2013212.","DOI":"10.1038\/s41592-023-02151-z"},{"key":"2075_CR29","doi-asserted-by":"crossref","unstructured":"Sengupta S, Chakrabarty S, Soni R. Is SAM 2 better than SAM in medical image segmentation? arXiv [eess.IV]. 2024.","DOI":"10.1117\/12.3047370"},{"key":"2075_CR30","doi-asserted-by":"crossref","unstructured":"Xia Y, Fripp J, Chandra SS, Schwarz R, Engstrom C, Crozier S. Automated bone segmentation from large field of view 3D MR images of the hip joint. Phys Med Biol. 2013;58:7375\u201390.","DOI":"10.1088\/0031-9155\/58\/20\/7375"},{"key":"2075_CR31","doi-asserted-by":"crossref","unstructured":"Heckelman LN, Soher BJ, Spritzer CE, Lewis BD, DeFrate LE. Design and validation of a semi-automatic bone segmentation algorithm from MRI to improve research efficiency. Sci Rep. 2022;12:7825.","DOI":"10.1038\/s41598-022-11785-6"},{"key":"2075_CR32","unstructured":"Yildiz Z, Chen Y, Mazurowski MA. SAM & SAM 2 in 3D slicer: SegmentWithSAM extension for annotating medical images. arXiv [eess.IV]. 2024."},{"key":"2075_CR33","doi-asserted-by":"crossref","unstructured":"Fedorov A, Beichel R, Kalpathy-Cramer J, Finet J, Fillion-Robin J-C, Pujol S, et al. 3D slicer as an image computing platform for the quantitative imaging network. Magn Reson Imag. 2012;30:1323\u201341.","DOI":"10.1016\/j.mri.2012.05.001"},{"key":"2075_CR34","unstructured":"Archit A, Freckmann L, Pape C. MedicoSAM: towards foundation models for medical image segmentation. arXiv [eess.IV]. 2025."},{"key":"2075_CR35","unstructured":"Zhu J, Hamdi A, Qi Y, Jin Y, Wu J. Medical SAM 2: segment medical images as video via segment anything Model 2. arXiv [cs.CV]. 2024."},{"key":"2075_CR36","doi-asserted-by":"crossref","unstructured":"Ma J, He Y, Li F, Han L, You C, Wang B. Segment anything in medical images. Nat Commun. 2024;15:654.","DOI":"10.1038\/s41467-024-44824-z"},{"key":"2075_CR37","unstructured":"Bai Y, Yun B, Chen Z, Yu Q, Xia Y, Wang Y. RevSAM2: prompt SAM2 for medical image segmentation via reverse-propagation without fine-tuning. arXiv [cs.CV]. 2024."},{"key":"2075_CR38","doi-asserted-by":"crossref","unstructured":"Lei W, Xu W, Li K, Zhang X, Zhang S. MedLSAM: localize and segment anything model for 3D CT images. Med Image Anal. 2025;99:103370.","DOI":"10.1016\/j.media.2024.103370"},{"key":"2075_CR39","unstructured":"Ma J, Yang Z, Kim S, Chen B, Baharoon M, Fallahpour A, et al. MedSAM2: segment anything in 3D medical images and videos. arXiv [eess.IV]. 2025."},{"key":"2075_CR40","doi-asserted-by":"crossref","unstructured":"Aleem S, Wang F, Maniparambil M, Arazo E, Dietlmeier J, Curran K, et al. Test-time adaptation with SaLIP: a cascade of SAM and CLIP for zero-shot medical image segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops. 2024. p. 5184\u201393.","DOI":"10.1109\/CVPRW63382.2024.00526"},{"key":"2075_CR41","doi-asserted-by":"crossref","unstructured":"Mattjie C, De Moura LV, Ravazio R, Kupssinsk\u00fc L, Parraga O, Delucis MM, et al. Zero-shot performance of the segment anything model (SAM) in 2D medical imaging: a comprehensive evaluation and practical guidelines. 2023 IEEE 23rd International Conference on Bioinformatics and Bioengineering (BIBE). IEEE; 2023. p. 108\u201312.","DOI":"10.1109\/BIBE60311.2023.00025"},{"key":"2075_CR42","unstructured":"Introducing SAM 2: the next generation of meta segment anything Model for videos and images. https:\/\/ai.meta.com\/blog\/segment-anything-2\/. Accessed 10 Apr 2025."}],"container-title":["BMC Medical Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s12880-025-02075-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s12880-025-02075-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s12880-025-02075-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T14:37:30Z","timestamp":1769524650000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1186\/s12880-025-02075-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,8]]},"references-count":44,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2026,12]]}},"alternative-id":["2075"],"URL":"https:\/\/doi.org\/10.1186\/s12880-025-02075-4","relation":{},"ISSN":["1471-2342"],"issn-type":[{"value":"1471-2342","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,8]]},"assertion":[{"value":"26 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This study used only pre-existing, publicly available datasets (listed in the Data availability section). Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"KB declares speaker fees from Canon Medical Systems Corporation and GE HealthCare. DT received honoraria for lectures by Bayer, GE, Roche, AstraZeneca, and Philips and holds shares in StratifAI GmbH, Germany and in Synagen GmbH, Germany. JNK declares ongoing consulting services for AstraZeneca, Panakeia, and Bioptimus. Furthermore, he holds shares in StratifAI, Synagen, and Spira Labs, has received an institutional research grant from GSK and AstraZeneca, as well as honoraria from AstraZeneca, Bayer, Daiichi Sankyo, Eisai, Janssen, Merck, MSD, BMS, Roche, Pfizer, and Fresenius.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"54"}}