{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:35:48Z","timestamp":1757619348856,"version":"3.44.0"},"publisher-location":"Singapore","reference-count":43,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819500352"},{"type":"electronic","value":"9789819500369"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-95-0036-9_26","type":"book-chapter","created":{"date-parts":[[2025,7,23]],"date-time":"2025-07-23T08:37:22Z","timestamp":1753259842000},"page":"305-316","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["FreqSAM2-UNet: Adapter Fine-Tuning Frequency-Aware Network of SAM2 for Universal Medical Segmentation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-3807-6128","authenticated-orcid":false,"given":"Chun","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0001-5925-2566","authenticated-orcid":false,"given":"Jingxing","family":"Cao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-3749-3118","authenticated-orcid":false,"given":"Yuxiao","family":"Gao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3080-6657","authenticated-orcid":false,"given":"Jianfeng","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,24]]},"reference":[{"issue":"9","key":"26_CR1","doi-asserted-by":"publisher","first-page":"1342","DOI":"10.1038\/s41591-018-0107-6","volume":"24","author":"J De Fauw","year":"2018","unstructured":"De Fauw, J.: Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 24(9), 1342\u20131350 (2018)","journal-title":"Nat. Med."},{"key":"26_CR2","doi-asserted-by":"publisher","unstructured":"Ronneberger, O.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015, LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"26_CR3","doi-asserted-by":"crossref","unstructured":"Kirillov, A.: Segment anything. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4015\u20134026. IEEE (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"26_CR4","unstructured":"Ravi, N.: SAM 2: segment anything in images and videos. arXiv preprint arXiv:2408.00714 (2024)"},{"key":"26_CR5","doi-asserted-by":"crossref","unstructured":"Chen, T.: SAM-Adapter: adapting segment anything in underperformed scenes. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3367\u20133375. IEEE (2023)","DOI":"10.1109\/ICCVW60793.2023.00361"},{"key":"26_CR6","doi-asserted-by":"crossref","unstructured":"Huang, D.: AlignSAM: aligning segment anything model to open context via reinforcement learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3205\u20133215. IEEE (2024)","DOI":"10.1109\/CVPR52733.2024.00309"},{"key":"26_CR7","unstructured":"Zhang, Y.: EVF-SAM: early vision-language fusion for text-prompted segment anything model. arXiv preprint arXiv:2406.20076, (2024)"},{"key":"26_CR8","doi-asserted-by":"publisher","unstructured":"Li, W.: TP-DRSeg: improving diabetic retinopathy lesion segmentation with explicit text-prompts assisted SAM. In: Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 743\u2013753. Springer Nature Switzerland, Cham (2024). https:\/\/doi.org\/10.1007\/978-3-031-72111-3_70","DOI":"10.1007\/978-3-031-72111-3_70"},{"key":"26_CR9","unstructured":"Zhang, R.: Personalize segment anything model with one shot. arXiv preprint arXiv:2305.03048 (2023)"},{"issue":"6","key":"26_CR10","doi-asserted-by":"publisher","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","volume":"39","author":"Z Zhou","year":"2019","unstructured":"Zhou, Z.: U-Net++: redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imag. 39(6), 1856\u20131867 (2019)","journal-title":"IEEE Trans. Med. Imag."},{"key":"26_CR11","doi-asserted-by":"publisher","unstructured":"Xiong, X.: Mammo-SAM: adapting foundation segment anything model for automatic breast mass segmentation in whole mammograms. In: Proceedings of the International Workshop on Machine Learning in Medical Imaging, pp. 176\u2013185. Springer Nature Switzerland (2023). https:\/\/doi.org\/10.1007\/978-3-031-45673-2_18","DOI":"10.1007\/978-3-031-45673-2_18"},{"key":"26_CR12","unstructured":"Dosovitskiy, A.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"26_CR13","unstructured":"Xiong, X.: SAM2-UNet: segment anything 2 makes strong encoder for natural and medical image segmentation. arXiv preprint arXiv:2408.08870 (2024)"},{"key":"26_CR14","unstructured":"Ryali, C.: Hiera: A hierarchical vision transformer without the bells-and-whistles. In: International Conference on Machine Learning, pp. 29441\u201329454. PMLR (2023)"},{"key":"26_CR15","doi-asserted-by":"crossref","unstructured":"Chen, L.: Frequency-aware feature fusion for dense image prediction. IEEE Trans. Pattern Anal. Mach. Intell. 46, 10763\u201310780 (2024)","DOI":"10.1109\/TPAMI.2024.3449959"},{"key":"26_CR16","doi-asserted-by":"publisher","unstructured":"Zhang, Y.: TransFuse: fusing transformers and CNNs for medical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2021: 24th International Conference, Strasbourg, France, September 27\u2013October 1, 2021, Proceedings, Part I, pp. 14\u201324. Springer International Publishing (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_2","DOI":"10.1007\/978-3-030-87193-2_2"},{"key":"26_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.dib.2019.104863","volume":"28","author":"W Al-Dhabyani","year":"2020","unstructured":"Al-Dhabyani, W.: Dataset of breast ultrasound images. Data Brief 28, 104863 (2020)","journal-title":"Data Brief"},{"key":"26_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.artmed.2020.101880","volume":"107","author":"MH Yap","year":"2020","unstructured":"Yap, M.H.: Breast ultrasound region of interest detection and lesion localisation. Artif. Intell. Med. 107, 101880 (2020)","journal-title":"Artif. Intell. Med."},{"issue":"8","key":"26_CR19","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0221535","volume":"14","author":"Z Zhuang","year":"2019","unstructured":"Zhuang, Z.: An RDAU-NET model for lesion segmentation in breast ultrasound images. PLoS ONE 14(8), e0221535 (2019)","journal-title":"PLoS ONE"},{"key":"26_CR20","doi-asserted-by":"publisher","unstructured":"Jha, D.: KVASIR-SEG: a segmented polyp dataset. In: MMM 2020, LNCS, vol. 11962, pp. 451\u2013462. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-37734-2_37","DOI":"10.1007\/978-3-030-37734-2_37"},{"key":"26_CR21","doi-asserted-by":"crossref","unstructured":"Bernal, J.: WM-DOVA maps for accurate polyp highlighting in colonoscopy: validation vs. saliency maps from physicians. Comput. Med. Imaging Graph. 43, 99\u2013111 (2015)","DOI":"10.1016\/j.compmedimag.2015.02.007"},{"issue":"2","key":"26_CR22","doi-asserted-by":"publisher","first-page":"630","DOI":"10.1109\/TMI.2015.2487997","volume":"35","author":"N Tajbakhsh","year":"2015","unstructured":"Tajbakhsh, N.: Automated polyp detection in colonoscopy videos using shape and context information. IEEE Trans. Med. Imaging 35(2), 630\u2013644 (2015)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"26_CR23","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1007\/s11548-013-0926-3","volume":"9","author":"J Silva","year":"2014","unstructured":"Silva, J.: Toward embedded detection of polyps in WCE images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 9, 283\u2013293 (2014)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"26_CR24","doi-asserted-by":"crossref","unstructured":"Ngoc Lan, P.: NeouNet: Towards accurate colon polyp segmentation and neoplasm detection. Adv. Vis. Comput. 16(II), 15\u201328 (2021)","DOI":"10.1007\/978-3-030-90436-4_2"},{"issue":"1","key":"26_CR25","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1038\/s41597-023-01981-y","volume":"10","author":"S Ali","year":"2023","unstructured":"Ali, S.: A multi-centre polyp detection and segmentation dataset for generalisability assessment. Sci. Data 10(1), 75 (2023)","journal-title":"Sci. Data"},{"key":"26_CR26","doi-asserted-by":"publisher","unstructured":"Wei, J.: Shallow attention network for polyp segmentation. In: de Bruijne, M., et al. (eds.) Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021, LNCS, vol. 12902, pp. 699\u2013708. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_66","DOI":"10.1007\/978-3-030-87193-2_66"},{"key":"26_CR27","doi-asserted-by":"publisher","unstructured":"Zhao, X.: Automatic polyp segmentation via multi-scale subtraction network. In: Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2021, LNCS, vol. 12902, pp. 120\u2013130. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_12","DOI":"10.1007\/978-3-030-87193-2_12"},{"key":"26_CR28","doi-asserted-by":"publisher","unstructured":"Nguyen-Mau, T.H.: PEFNet: positional embedding feature for polyp segmentation. In: International Conference on Multimedia Modeling, LNCS, vol. 13361, pp. 240\u2013251. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-27818-1_20","DOI":"10.1007\/978-3-031-27818-1_20"},{"key":"26_CR29","doi-asserted-by":"crossref","unstructured":"Trinh, Q.H.: M2UNet: MetaFormer Multi-Scale upsampling network for polyp segmentation. In: 31st European Signal Processing Conference (EUSIPCO 2023), pp. 1115\u20131119. IEEE (2023)","DOI":"10.23919\/EUSIPCO58844.2023.10290110"},{"key":"26_CR30","doi-asserted-by":"crossref","unstructured":"Bui, N.T.: MEGANet: Multi-scale edge-guided attention network for weak boundary polyp segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 7985\u20137994. IEEE (2024)","DOI":"10.1109\/WACV57701.2024.00780"},{"issue":"1","key":"26_CR31","doi-asserted-by":"publisher","first-page":"9803","DOI":"10.1038\/s41598-023-36940-5","volume":"13","author":"RG Dumitru","year":"2023","unstructured":"Dumitru, R.G.: Using DUCK-Net for polyp image segmentation. Sci. Rep. 13(1), 9803 (2023)","journal-title":"Sci. Rep."},{"issue":"9","key":"26_CR32","doi-asserted-by":"publisher","first-page":"2273","DOI":"10.1109\/TMI.2022.3162111","volume":"41","author":"J Song","year":"2022","unstructured":"Song, J.: Global and local feature reconstruction for medical image segmentation. IEEE Trans. Med. Imag. 41(9), 2273\u20132284 (2022)","journal-title":"IEEE Trans. Med. Imag."},{"key":"26_CR33","doi-asserted-by":"publisher","unstructured":"Wang, J.: Stepwise feature fusion: Local guides global. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2022, LNCS, vol. 13433, pp. 110\u2013120. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16437-8_11","DOI":"10.1007\/978-3-031-16437-8_11"},{"key":"26_CR34","doi-asserted-by":"crossref","unstructured":"Trinh, Q.H.: Meta-Polyp: a baseline for efficient polyp segmentation. In: 2023 IEEE 36th International Symposium on Computer-Based Medical Systems (CBMS), pp. 742\u2013747. IEEE (2023)","DOI":"10.1109\/CBMS58004.2023.00312"},{"key":"26_CR35","doi-asserted-by":"crossref","unstructured":"Rahman, M.M.: Medical image segmentation via cascaded attention decoding. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 6222\u20136231. IEEE (2023)","DOI":"10.1109\/WACV56688.2023.00616"},{"key":"26_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103288","volume":"97","author":"L Cai","year":"2024","unstructured":"Cai, L.: Know your orientation: a viewpoint-aware framework for polyp segmentation. Med. Image Anal. 97, 103288 (2024)","journal-title":"Med. Image Anal."},{"key":"26_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2021.103299","volume":"72","author":"Y Yan","year":"2022","unstructured":"Yan, Y.: Accurate segmentation of breast tumors using AE U-net with HDC model in ultrasound images. Biomed. Signal Process. Control 72, 103299 (2022)","journal-title":"Biomed. Signal Process. Control"},{"key":"26_CR38","doi-asserted-by":"crossref","unstructured":"Abraham, N.: A novel focal tversky loss function with improved attention u-net for lesion segmentation. In: 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), pp. 683\u2013687. IEEE (2019)","DOI":"10.1109\/ISBI.2019.8759329"},{"key":"26_CR39","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A.: UNETR: Transformers for 3D medical image segmentation. In: Proceedings of the IEEE Winter Conference on Applications of Computer Vision, pp. 574\u2013584. IEEE (2022)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"26_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2020.102027","volume":"61","author":"M Byra","year":"2020","unstructured":"Byra, M.: Breast mass segmentation in ultrasound with selective kernel U-Net convolutional neural network. Biomed. Signal Process. Control 61, 102027 (2020)","journal-title":"Biomed. Signal Process. Control"},{"key":"26_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.104425","volume":"81","author":"Y Lyu","year":"2023","unstructured":"Lyu, Y.: AMS-PAN: breast ultrasound image segmentation model combining attention mechanism and multi-scale features. Biomed. Signal Process. Control 81, 104425 (2023)","journal-title":"Biomed. Signal Process. Control"},{"key":"26_CR42","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.123265","volume":"246","author":"G Chen","year":"2024","unstructured":"Chen, G.: ESKNet: an enhanced adaptive selection kernel convolution for ultrasound breast tumors segmentation. Expert Syst. Appl. 246, 123265 (2024)","journal-title":"Expert Syst. Appl."},{"key":"26_CR43","doi-asserted-by":"crossref","unstructured":"Liu, S.: Receptive field block net for accurate and fast object detection. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 385\u2013400 (2018)","DOI":"10.1007\/978-3-030-01252-6_24"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-0036-9_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T19:48:35Z","timestamp":1757274515000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-0036-9_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819500352","9789819500369"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-0036-9_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"24 July 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ningbo","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 July 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 July 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/icg\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}