{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T16:47:57Z","timestamp":1764175677951,"version":"3.40.3"},"publisher-location":"Cham","reference-count":22,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031665349"},{"type":"electronic","value":"9783031665356"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-66535-6_15","type":"book-chapter","created":{"date-parts":[[2024,7,26]],"date-time":"2024-07-26T12:03:45Z","timestamp":1721995425000},"page":"131-139","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["TinySAM-Med3D: A Lightweight Segment Anything Model for\u00a0Volumetric Medical Imaging with\u00a0Mixture of\u00a0Experts"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-6909-8080","authenticated-orcid":false,"given":"Tianyuan","family":"Song","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4039-4505","authenticated-orcid":false,"given":"Guixia","family":"Kang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7866-3339","authenticated-orcid":false,"given":"Yiqing","family":"Shen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,25]]},"reference":[{"key":"15_CR1","unstructured":"Cheng, J., et\u00a0al.: SAM-Med2D. arXiv preprint arXiv:2308.16184 (2023)"},{"key":"15_CR2","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16\u00a0$$\\times $$\u00a016 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"15_CR3","unstructured":"Gao, H., Li, Y., Long, K., Yang, M., Shen, Y.: A survey for foundation models in autonomous driving. arXiv preprint arXiv:2402.01105 (2024)"},{"key":"15_CR4","doi-asserted-by":"publisher","unstructured":"Hatamizadeh, A., Nath, V., Tang, Y., Yang, D., Roth, H.R., Xu, D.: Swin UNETR: Swin transformers for semantic segmentation of brain tumors in MRI images. In: Crimi, A., Bakas, S. (eds.) International MICCAI Brainlesion Workshop. LNCS, vol. 12962, pp. 272\u2013284. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-031-08999-2_22","DOI":"10.1007\/978-3-031-08999-2_22"},{"key":"15_CR5","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"15_CR6","unstructured":"Howard, A., Zhmoginov, A., Chen, L.C., Sandler, M., Zhu, M.: Inverted residuals and linear bottlenecks: mobile networks for classification, detection and segmentation (2018)"},{"issue":"1","key":"15_CR7","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1162\/neco.1991.3.1.79","volume":"3","author":"RA Jacobs","year":"1991","unstructured":"Jacobs, R.A., Jordan, M.I., Nowlan, S.J., Hinton, G.E.: Adaptive mixtures of local experts. Neural Comput. 3(1), 79\u201387 (1991)","journal-title":"Neural Comput."},{"key":"15_CR8","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: Towards the unification of generative and discriminative visual foundation model: a survey. arXiv preprint arXiv:2312.10163 (2023)","DOI":"10.1007\/s00371-024-03608-8"},{"issue":"1","key":"15_CR10","doi-asserted-by":"publisher","first-page":"654","DOI":"10.1038\/s41467-024-44824-z","volume":"15","author":"J Ma","year":"2024","unstructured":"Ma, J., He, Y., Li, F., Han, L., You, C., Wang, B.: Segment anything in medical images. Nat. Commun. 15(1), 654 (2024)","journal-title":"Nat. Commun."},{"key":"15_CR11","unstructured":"Puigcerver, J., Riquelme, C., Mustafa, B., Houlsby, N.: From sparse to soft mixtures of experts. arXiv preprint arXiv:2308.00951 (2023)"},{"key":"15_CR12","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W., Frangi, A. (eds.) Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, 5\u20139 October 2015, Proceedings, Part III 18, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"15_CR13","doi-asserted-by":"publisher","unstructured":"Roy, S., et al.: MedNeXt: transformer-driven scaling of convnets for medical image segmentation. In: Greenspan, H., et al. (eds.) International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 405\u2013415. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-43901-8_39","DOI":"10.1007\/978-3-031-43901-8_39"},{"key":"15_CR14","doi-asserted-by":"crossref","unstructured":"Shen, Y., Xu, L., Yang, Y., Li, Y., Guo, Y.: Self-distillation from the last mini-batch for consistency regularization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11943\u201311952 (2022)","DOI":"10.1109\/CVPR52688.2022.01164"},{"key":"15_CR15","doi-asserted-by":"crossref","unstructured":"Shen, Y., Xu, L., Yang, Y., Li, Y., Guo, Y.: Mixed sample augmentation for online distillation. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.\u00a01\u20135. IEEE (2023)","DOI":"10.1109\/ICASSP49357.2023.10096043"},{"key":"15_CR16","unstructured":"Shu, H., et al.: TinySAM: pushing the envelope for efficient segment anything model. arXiv preprint arXiv:2312.13789 (2023)"},{"key":"15_CR17","unstructured":"Wang, C., et al.: SAM-Med: a medical image annotation framework based on large vision model. arXiv preprint arXiv:2307.05617 (2023)"},{"key":"15_CR18","unstructured":"Wang, H., et\u00a0al.: SAM-Med3D. arXiv preprint arXiv:2310.15161 (2023)"},{"key":"15_CR19","doi-asserted-by":"publisher","unstructured":"Wu, K., et al.: TinyVit: fast pretraining distillation for small vision transformers. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) European Conference on Computer Vision, pp. 68\u201385. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19803-8_5","DOI":"10.1007\/978-3-031-19803-8_5"},{"key":"15_CR20","doi-asserted-by":"crossref","unstructured":"Xiong, Y., et\u00a0al.: EfficientSAM: leveraged masked image pretraining for efficient segment anything. arXiv preprint arXiv:2312.00863 (2023)","DOI":"10.1109\/CVPR52733.2024.01525"},{"key":"15_CR21","doi-asserted-by":"crossref","unstructured":"Xue, F., Shi, Z., Wei, F., Lou, Y., Liu, Y., You, Y.: Go wider instead of deeper. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a036, pp. 8779\u20138787 (2022)","DOI":"10.1609\/aaai.v36i8.20858"},{"key":"15_CR22","unstructured":"Zhang, C., et al.: Faster segment anything: towards lightweight SAM for mobile applications. arXiv preprint arXiv:2306.14289 (2023)"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-66535-6_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T21:53:58Z","timestamp":1732485238000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-66535-6_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031665349","9783031665356"],"references-count":22,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-66535-6_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"25 July 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"AIME","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Intelligence in Medicine","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Salt Lake City, UT","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 July 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 July 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"aime2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/aime24.aimedicine.info\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}