{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T02:52:57Z","timestamp":1774234377648,"version":"3.50.1"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"13","license":[{"start":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T00:00:00Z","timestamp":1752451200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T00:00:00Z","timestamp":1752451200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100021171","name":"Guangdong Basic and Applied Basic Research Foundation","doi-asserted-by":"crossref","award":["2024A1515011825"],"award-info":[{"award-number":["2024A1515011825"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100021171","name":"Guangdong Basic and Applied Basic Research Foundation","doi-asserted-by":"crossref","award":["2022A1515110696, 2025A1515012945, 2025A1515012911"],"award-info":[{"award-number":["2022A1515110696, 2025A1515012945, 2025A1515012911"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100021171","name":"Guangdong Basic and Applied Basic Research Foundation","doi-asserted-by":"crossref","award":["2021A1515011091"],"award-info":[{"award-number":["2021A1515011091"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Doctor Starting Fund of Hanshan Normal University, China","award":["QD202324"],"award-info":[{"award-number":["QD202324"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["12305409"],"award-info":[{"award-number":["12305409"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Shenzhen Science and Technology Program of China","award":["JCYJ20230807145004008, JCYJ20240813155826035"],"award-info":[{"award-number":["JCYJ20230807145004008, JCYJ20240813155826035"]}]},{"name":"Research Matching Grant Scheme from the University Grants Council of the Hong Kong Special Administrative Region, China","award":["No. 2024\/3003"],"award-info":[{"award-number":["No. 2024\/3003"]}]},{"DOI":"10.13039\/501100010226","name":"Educational Commission of Guangdong Province","doi-asserted-by":"crossref","award":["2020ZDZX3056"],"award-info":[{"award-number":["2020ZDZX3056"]}],"id":[{"id":"10.13039\/501100010226","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s00371-025-04092-4","type":"journal-article","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T05:57:38Z","timestamp":1752472658000},"page":"11139-11149","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["ZAP-2.5DSAM: zero additional parameters advancing 2.5D SAM adaptation to 3D tumor segmentation"],"prefix":"10.1007","volume":"41","author":[{"given":"Cai","family":"Guo","sequence":"first","affiliation":[]},{"given":"Yuxi","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Bishenghui","family":"Tao","sequence":"additional","affiliation":[]},{"given":"Jianzhong","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hong-Ning","family":"Dai","sequence":"additional","affiliation":[]},{"given":"Ping","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,14]]},"reference":[{"issue":"1","key":"4092_CR1","first-page":"4128","volume":"13","author":"M Antonelli","year":"2022","unstructured":"Antonelli, M., Reinke, A., Bakas, S., Farahani, K., Kopp-Schneider, A., Landman, B.A., Litjens, G., Menze, B., Ronneberger, O., Summers, R.M., et al.: Nature communications. Med. Segm. Decathlon 13(1), 4128 (2022)","journal-title":"Med. Segm. Decathlon"},{"key":"4092_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102680","volume":"84","author":"P Bilic","year":"2023","unstructured":"Bilic, P., Christ, P., Li, H.B., Vorontsov, E., Ben-Cohen, A., Kaissis, G., Szeskin, A., Jacobs, C., Mamani, G.E.H., Chartrand, G., et al.: The liver tumor segmentation benchmark (lits). Med. Image Anal. 84, 102,680 (2023)","journal-title":"Med. Image Anal."},{"key":"4092_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103310","volume":"98","author":"C Chen","year":"2024","unstructured":"Chen, C., Miao, J., Wu, D., Zhong, A., Yan, Z., Kim, S., Hu, J., Liu, Z., Sun, L., Li, X., et al.: Ma-sam: modality-agnostic sam adaptation for 3d medical image segmentation. Med. Image Anal. 98, 103, 310 (2024)","journal-title":"Med. Image Anal."},{"key":"4092_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103280","volume":"97","author":"J Chen","year":"2024","unstructured":"Chen, J., Mei, J., Li, X., Lu, Y., Yu, Q., Wei, Q., Luo, X., Xie, Y., Adeli, E., Wang, Y., et al.: Transunet: rethinking the u-net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 97, 103280 (2024)","journal-title":"Med. Image Anal."},{"issue":"1","key":"4092_CR5","doi-asserted-by":"publisher","first-page":"15497","DOI":"10.1038\/s41598-018-33860-7","volume":"8","author":"G Chlebus","year":"2018","unstructured":"Chlebus, G., Schenk, A., Moltz, J.H., van Ginneken, B., Hahn, H.K., Meine, H.: Automatic liver tumor segmentation in ct with fully convolutional neural networks and object-based postprocessing. Sci. Rep. 8(1), 15497 (2018)","journal-title":"Sci. Rep."},{"key":"4092_CR6","unstructured":"Dosovitskiy, A.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"4092_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103324","volume":"98","author":"S Gong","year":"2024","unstructured":"Gong, S., Zhong, Y., Ma, W., Li, J., Wang, Z., Zhang, J., Heng, P.A., Dou, Q.: 3dsam-adapter: holistic adaptation of sam from 2d to 3d for promptable tumor segmentation. Med. Image Anal. 98, 103,324 (2024)","journal-title":"Med. Image Anal."},{"key":"4092_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2020.101821","volume":"67","author":"N Heller","year":"2021","unstructured":"Heller, N., Isensee, F., Maier-Hein, K.H., Hou, X., Xie, C., Li, F., Nan, Y., Mu, G., Lin, Z., Han, M., et al.: The state of the art in kidney and kidney tumor segmentation in contrast-enhanced ct imaging: results of the kits19 challenge. Med. Image Anal. 67, 101821 (2021)","journal-title":"Med. Image Anal."},{"key":"4092_CR9","unstructured":"Heller, N., Isensee, F., Trofimova, D., Tejpaul, R., Zhao, Z., Chen, H., Wang, L., Golts, A., Khapun, D., Shats, D., et\u00a0al.: The kits21 challenge: Automatic segmentation of kidneys, renal tumors, and renal cysts in corticomedullary-phase ct. arXiv preprint arXiv:2307.01984 (2023)"},{"issue":"2","key":"4092_CR10","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"4092_CR11","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.Y., et\u00a0al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"4092_CR12","unstructured":"Lee, H.H., Bao, S., Huo, Y., Landman, B.A.: 3d ux-net: a large kernel volumetric convnet modernizing hierarchical transformer for medical image segmentation. arXiv preprint arXiv:2209.15076 (2022)"},{"key":"4092_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103370","volume":"99","author":"W Lei","year":"2025","unstructured":"Lei, W., Xu, W., Li, K., Zhang, X., Zhang, S.: Medlsam: localize and segment anything model for 3d ct images. Med. Image Anal. 99, 103,370 (2025)","journal-title":"Med. Image Anal."},{"key":"4092_CR14","doi-asserted-by":"crossref","unstructured":"Li, H., Liu, H., Hu, D., Wang, J., Oguz, I.: Promise: prompt-driven 3d medical image segmentation using pretrained image foundation models. In: 2024 IEEE International Symposium on Biomedical Imaging (ISBI), pp. 1\u20135. IEEE (2024)","DOI":"10.1109\/ISBI56570.2024.10635207"},{"key":"4092_CR15","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.media.2017.07.005","volume":"42","author":"G Litjens","year":"2017","unstructured":"Litjens, G., Kooi, T., Bejnordi, B.E., Setio, A.A.A., Ciompi, F., Ghafoorian, M., Van Der Laak, J.A., Van Ginneken, B., S\u00e1nchez, C.I.: A survey on deep learning in medical image analysis. Med. image Anal. 42, 60\u201388 (2017)","journal-title":"Med. image Anal."},{"key":"4092_CR16","unstructured":"Loshchilov, I.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"issue":"1","key":"4092_CR17","doi-asserted-by":"publisher","first-page":"654","DOI":"10.1038\/s41467-024-44824-z","volume":"15","author":"J Ma","year":"2024","unstructured":"Ma, J., He, Y., Li, F., Han, L., You, C., Wang, B.: Segment anything in medical images. Nat. Commun. 15(1), 654 (2024)","journal-title":"Nat. Commun."},{"key":"4092_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102918","volume":"89","author":"MA Mazurowski","year":"2023","unstructured":"Mazurowski, M.A., Dong, H., Gu, H., Yang, J., Konz, N., Zhang, Y.: Segment anything model for medical image analysis: an experimental study. Med. Image Anal. 89, 102918 (2023)","journal-title":"Med. Image Anal."},{"key":"4092_CR19","doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., Ahmadi, S.A.: V-net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 4th international conference on 3D vision (3DV), pp. 565\u2013571. Ieee (2016)","DOI":"10.1109\/3DV.2016.79"},{"key":"4092_CR20","unstructured":"Nikolov, S., Blackwell, S., Zverovitch, A., Mendes, R., Livne, M., De\u00a0Fauw, J., Patel, Y., Meyer, C., Askham, H., Romera-Paredes, B., et\u00a0al.: Deep learning to achieve clinically applicable segmentation of head and neck anatomy for radiotherapy. arXiv preprint arXiv:1809.04430 (2018)"},{"key":"4092_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2020.104003","volume":"126","author":"I Pacal","year":"2020","unstructured":"Pacal, I., Karaboga, D., Basturk, A., Akay, B., Nalbantoglu, U.: A comprehensive review of deep learning in colon cancer. Comput. Biol. Med. 126, 104003 (2020)","journal-title":"Comput. Biol. Med."},{"key":"4092_CR22","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical image computing and computer-assisted intervention\u2013MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18, pp. 234\u2013241. Springer (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"4092_CR23","doi-asserted-by":"crossref","unstructured":"Shaker, A.M., Maaz, M., Rasheed, H., Khan, S., Yang, M.H., Khan, F.S.: Unetr++: delving into efficient and accurate 3d medical image segmentation. IEEE Transactions on Medical Imaging (2024)","DOI":"10.1109\/TMI.2024.3398728"},{"key":"4092_CR24","doi-asserted-by":"publisher","first-page":"82031","DOI":"10.1109\/ACCESS.2021.3086020","volume":"9","author":"N Siddique","year":"2021","unstructured":"Siddique, N., Paheding, S., Elkin, C.P., Devabhaktuni, V.: U-net and its variants for medical image segmentation: a review of theory and applications. IEEE Access 9, 82031\u201382057 (2021)","journal-title":"IEEE Access"},{"key":"4092_CR25","doi-asserted-by":"crossref","unstructured":"Tang, Y., Yang, D., Li, W., Roth, H.R., Landman, B., Xu, D., Nath, V., Hatamizadeh, A.: Self-supervised pre-training of swin transformers for 3d medical image analysis. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 20,730\u201320,740 (2022)","DOI":"10.1109\/CVPR52688.2022.02007"},{"key":"4092_CR26","unstructured":"Wald, T., Roy, S., Koehler, G., Disch, N., Rokuss, M.R., Holzschuh, J., Zimmerer, D., Maier-Hein, K.: Sam. md: zero-shot medical image segmentation capabilities of the segment anything model. In: Medical Imaging with Deep Learning, short paper track (2023)"},{"key":"4092_CR27","doi-asserted-by":"crossref","unstructured":"Wang, G., Ye, J., Cheng, J., Li, T., Chen, Z., Cai, J., He, J., Zhuang, B.: Sam-med3d-moe: towards a non-forgetting segment anything model via mixture of experts for 3d medical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 552\u2013561. Springer (2024)","DOI":"10.1007\/978-3-031-72114-4_53"},{"key":"4092_CR28","doi-asserted-by":"crossref","unstructured":"Wenxuan, W., Chen, C., Meng, D., Hong, Y., Sen, Z., Jiangyun, L.: Transbts: multimodal brain tumor segmentation using transformer. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, pp. 109\u2013119 (2021)","DOI":"10.1007\/978-3-030-87193-2_11"},{"key":"4092_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Shen, Z., Jiao, R.: Segment anything model for medical image segmentation: current applications and future directions. Comput. Biol. Med. pp. 108238 (2024)","DOI":"10.1016\/j.compbiomed.2024.108238"},{"key":"4092_CR30","unstructured":"Zhou, H.Y., Guo, J., Zhang, Y., Yu, L., Wang, L., Yu, Y.: nnformer: interleaved transformer for volumetric segmentation. arXiv preprint arXiv:2109.03201 (2021)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04092-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04092-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04092-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T14:01:05Z","timestamp":1758722465000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04092-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,14]]},"references-count":30,"journal-issue":{"issue":"13","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["4092"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04092-4","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-6916638\/v1","asserted-by":"object"}]},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,14]]},"assertion":[{"value":"24 June 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 July 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}