{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,5]],"date-time":"2026-03-05T15:30:09Z","timestamp":1772724609246,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,2,26]],"date-time":"2025-02-26T00:00:00Z","timestamp":1740528000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,2,26]],"date-time":"2025-02-26T00:00:00Z","timestamp":1740528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"The Science and Technology Development Fund, Macau SAR","award":["0074\/2022\/A"],"award-info":[{"award-number":["0074\/2022\/A"]}]},{"name":"The Science and Technology Development Fund, Macau SAR","award":["0074\/2022\/A"],"award-info":[{"award-number":["0074\/2022\/A"]}]},{"name":"The Science and Technology Development Fund, Macau SAR","award":["0074\/2022\/A"],"award-info":[{"award-number":["0074\/2022\/A"]}]},{"name":"The Science and Technology Development Fund, Macau SAR","award":["0074\/2022\/A"],"award-info":[{"award-number":["0074\/2022\/A"]}]},{"name":"The Science and Technology Development Fund, Macau SAR","award":["0074\/2022\/A"],"award-info":[{"award-number":["0074\/2022\/A"]}]},{"DOI":"10.13039\/501100004733","name":"University of Macau","doi-asserted-by":"crossref","award":["MYRG-GRG2023-00061-FST-UMDF"],"award-info":[{"award-number":["MYRG-GRG2023-00061-FST-UMDF"]}],"id":[{"id":"10.13039\/501100004733","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100004733","name":"University of Macau","doi-asserted-by":"crossref","award":["MYRG-GRG2023-00061-FST-UMDF"],"award-info":[{"award-number":["MYRG-GRG2023-00061-FST-UMDF"]}],"id":[{"id":"10.13039\/501100004733","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100004733","name":"University of Macau","doi-asserted-by":"crossref","award":["MYRG-GRG2023-00061-FST-UMDF"],"award-info":[{"award-number":["MYRG-GRG2023-00061-FST-UMDF"]}],"id":[{"id":"10.13039\/501100004733","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100004733","name":"University of Macau","doi-asserted-by":"crossref","award":["MYRG-GRG2023-00061-FST-UMDF"],"award-info":[{"award-number":["MYRG-GRG2023-00061-FST-UMDF"]}],"id":[{"id":"10.13039\/501100004733","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100004733","name":"University of Macau","doi-asserted-by":"crossref","award":["MYRG-GRG2023-00061-FST-UMDF"],"award-info":[{"award-number":["MYRG-GRG2023-00061-FST-UMDF"]}],"id":[{"id":"10.13039\/501100004733","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100015805","name":"Shenzhen Municipal Science and Technology Innovation Council","doi-asserted-by":"publisher","award":["SGDX20220530111001006"],"award-info":[{"award-number":["SGDX20220530111001006"]}],"id":[{"id":"10.13039\/501100015805","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100015805","name":"Shenzhen Municipal Science and Technology Innovation Council","doi-asserted-by":"publisher","award":["SGDX20220530111001006"],"award-info":[{"award-number":["SGDX20220530111001006"]}],"id":[{"id":"10.13039\/501100015805","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100015805","name":"Shenzhen Municipal Science and Technology Innovation Council","doi-asserted-by":"publisher","award":["SGDX20220530111001006"],"award-info":[{"award-number":["SGDX20220530111001006"]}],"id":[{"id":"10.13039\/501100015805","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100015805","name":"Shenzhen Municipal Science and Technology Innovation Council","doi-asserted-by":"publisher","award":["SGDX20220530111001006"],"award-info":[{"award-number":["SGDX20220530111001006"]}],"id":[{"id":"10.13039\/501100015805","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100015805","name":"Shenzhen Municipal Science and Technology Innovation Council","doi-asserted-by":"publisher","award":["SGDX20220530111001006"],"award-info":[{"award-number":["SGDX20220530111001006"]}],"id":[{"id":"10.13039\/501100015805","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s00371-025-03809-9","type":"journal-article","created":{"date-parts":[[2025,2,26]],"date-time":"2025-02-26T06:46:48Z","timestamp":1740552408000},"page":"7351-7365","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Learning few-shot semantic segmentation with error-filtered segment anything model"],"prefix":"10.1007","volume":"41","author":[{"given":"Chen-Bin","family":"Feng","sequence":"first","affiliation":[]},{"given":"Qi","family":"Lai","sequence":"additional","affiliation":[]},{"given":"Kangdao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Houcheng","family":"Su","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Kaixi","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Chi-Man","family":"Vong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,26]]},"reference":[{"key":"3809_CR1","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2021","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"3809_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110962","volume":"280","author":"J Yang","year":"2023","unstructured":"Yang, J., Ke, A., Yu, Y., Cai, B.: Scene sketch semantic segmentation with hierarchical transformer. Knowl. Based Syst. 280, 110,962 (2023)","journal-title":"Knowl. Based Syst."},{"key":"3809_CR3","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-024-03569-y","author":"L Li","year":"2024","unstructured":"Li, L., Ding, J., Cui, H., Chen, Z., Liao, G.: LiteMSNet: a lightweight semantic segmentation network with multi-scale feature extraction for urban streetscape scenes. Vis Comput (2024). https:\/\/doi.org\/10.1007\/s00371-024-03569-y","journal-title":"Vis Comput"},{"key":"3809_CR4","doi-asserted-by":"crossref","unstructured":"Shaban, A., Bansal, S., Liu, Z., Essa, I., Boots, B.: One-shot learning for semantic segmentation (2017). arXiv preprint arXiv:1709.03410","DOI":"10.5244\/C.31.167"},{"key":"3809_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110887","volume":"278","author":"Q Li","year":"2023","unstructured":"Li, Q., Sun, B., Bhanu, B.: Lite-FENet: lightweight multi-scale feature enrichment network for few-shot segmentation. Knowl. Based Syst. 278, 110,887 (2023)","journal-title":"Knowl. Based Syst."},{"key":"3809_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110698","volume":"275","author":"Y Yang","year":"2023","unstructured":"Yang, Y., Chen, Q., Liu, Q.: A dual-channel network for cross-domain one-shot semantic segmentation via adversarial learning. Knowl. Based Syst. 275, 110698 (2023)","journal-title":"Knowl. Based Syst."},{"key":"3809_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110520","volume":"270","author":"J Hu","year":"2023","unstructured":"Hu, J., Gu, X., Wang, Z., Gu, X.: Mixture of calibrated networks for domain generalization in brain tumor segmentation. Knowl. Based Syst. 270, 110,520 (2023)","journal-title":"Knowl. Based Syst."},{"key":"3809_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110987","volume":"280","author":"Z Liang","year":"2023","unstructured":"Liang, Z., Zhao, K., Liang, G., Li, S., Wu, Y., Zhou, Y.: MAXFormer: enhanced transformer for medical image segmentation with multi-attention and multi-scale features fusion. Knowl. Based Syst. 280, 110,987 (2023)","journal-title":"Knowl. Based Syst."},{"key":"3809_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.111348","volume":"285","author":"Y Feng","year":"2024","unstructured":"Feng, Y., Cao, Y., An, D., Liu, P., Liao, X., Yu, B.: Daunet: a u-shaped network combining deep supervision and attention for brain tumor segmentation. Knowl. Based Syst. 285, 111,348 (2024)","journal-title":"Knowl. Based Syst."},{"key":"3809_CR10","doi-asserted-by":"crossref","unstructured":"Chu, T., Chen, J., Sun, J., Lian, S., Wang, Z., Zuo, Z., Zhao, L., Xing, W., Lu, D.: Rethinking fast fourier convolution in image inpainting. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 23195\u201323205 (2023)","DOI":"10.1109\/ICCV51070.2023.02120"},{"key":"3809_CR11","doi-asserted-by":"crossref","unstructured":"Ko, K., Kim, C.S.: Continuously masked transformer for image inpainting. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13169\u201313178 (2023)","DOI":"10.1109\/ICCV51070.2023.01211"},{"key":"3809_CR12","doi-asserted-by":"crossref","unstructured":"Chen, X., Huang, L., Liu, Y., Shen, Y., Zhao, D., Zhao, H.: Anydoor: zero-shot object-level image customization (2023). arXiv preprint arXiv:2307.09481","DOI":"10.1109\/CVPR52733.2024.00630"},{"key":"3809_CR13","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.Y., et\u00a0al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"3809_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Y., Liu, N., Cao, Q., Yao, X., Han, J., Shao, L.: Learning non-target knowledge for few-shot semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11573\u201311582 (2022)","DOI":"10.1109\/CVPR52688.2022.01129"},{"key":"3809_CR15","doi-asserted-by":"crossref","unstructured":"Liu, J., Bao, Y., Xie, G.S., Xiong, H., Sonke, J.J., Gavves, E.: Dynamic prototype convolution network for few-shot semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11553\u201311562 (2022)","DOI":"10.1109\/CVPR52688.2022.01126"},{"key":"3809_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109018","volume":"133","author":"H Ding","year":"2023","unstructured":"Ding, H., Zhang, H., Jiang, X.: Self-regularized prototypical network for few-shot semantic segmentation. Pattern Recogn. 133, 109,018 (2023)","journal-title":"Pattern Recogn."},{"key":"3809_CR17","doi-asserted-by":"crossref","unstructured":"Min, J., Kang, D., Cho, M.: Hypercorrelation squeeze for few-shot segmentation (2021). arXiv preprint arXiv:2104.01538","DOI":"10.1109\/ICCV48922.2021.00686"},{"key":"3809_CR18","doi-asserted-by":"crossref","unstructured":"Peng, B., Tian, Z., Wu, X., Wang, C., Liu, S., Su, J., Jia, J.: Hierarchical dense correlation distillation for few-shot segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 23641\u201323651 (2023)","DOI":"10.1109\/CVPR52729.2023.02264"},{"key":"3809_CR19","doi-asserted-by":"crossref","unstructured":"Xu, Q., Zhao, W., Lin, G., Long, C.: Self-calibrated cross attention network for few-shot segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 655\u2013665 (2023)","DOI":"10.1109\/ICCV51070.2023.00067"},{"key":"3809_CR20","doi-asserted-by":"publisher","first-page":"8580","DOI":"10.1109\/TMM.2023.3238521","volume":"25","author":"H Liu","year":"2023","unstructured":"Liu, H., Peng, P., Chen, T., Wang, Q., Yao, Y., Hua, X.S.: FECANet: boosting few-shot semantic segmentation with feature-enhanced context-aware network. IEEE Trans. Multimedia 25, 8580\u20138592 (2023)","journal-title":"IEEE Trans. Multimedia"},{"key":"3809_CR21","doi-asserted-by":"crossref","unstructured":"Cao, Q., Chen, Y., Ma, C., Yang, X.: Few-shot rotation-invariant aerial image semantic segmentation. arXiv preprint arXiv:2306.11734 (2023)","DOI":"10.1109\/TGRS.2023.3338699"},{"key":"3809_CR22","unstructured":"Wu, J., Fu, R., Fang, H., Liu, Y., Wang, Z., Xu, Y., Jin, Y., Arbel, T.: Medical SAM adapter: adapting segment anything model for medical image segmentation (2023). arXiv preprint arXiv:2304.12620"},{"key":"3809_CR23","unstructured":"Zhou, T., Zhang, Y., Zhou, Y., Wu, Y., Gong, C.: Can sam segment polyps? (2023). arXiv preprint arXiv:2304.07583"},{"key":"3809_CR24","unstructured":"Yang, J., Gao, M., Li, Z., Gao, S., Wang, F., Zheng, F.: Track anything: segment anything meets videos (2023). arXiv preprint arXiv:2304.11968"},{"key":"3809_CR25","unstructured":"Cheng, Y., Li, L., Xu, Y., Li, X., Yang, Z., Wang, W., Yang, Y.: Segment and track anything (2023). arXiv preprint arXiv:2305.06558"},{"key":"3809_CR26","unstructured":"Yu, T., Feng, R., Feng, R., Liu, J., Jin, X., Zeng, W., Chen, Z.: Inpaint anything: segment anything meets image inpainting (2023). arXiv preprint arXiv:2304.06790"},{"key":"3809_CR27","doi-asserted-by":"crossref","unstructured":"Yuan, H., Li, X., Zhou, C., Li, Y., Chen, K., Loy, C.C.: Open-vocabulary SAM: segment and recognize twenty-thousand classes interactively (2024). arXiv preprint arXiv:2401.02955","DOI":"10.1007\/978-3-031-72775-7_24"},{"key":"3809_CR28","unstructured":"Jia, C., Yang, Y., Xia, Y., Chen, Y.T., Parekh, Z., Pham, H., Le, Q., Sung, Y.H., Li, Z., Duerig, T.: Scaling up visual and vision-language representation learning with noisy text supervision. In: International Conference on Machine Learning, PMLR. pp. 4904\u20134916 (2021)"},{"key":"3809_CR29","doi-asserted-by":"crossref","unstructured":"Ye, M., Zhang, J., Liu, J., Liu, C., Yin, B., Liu, C., Du, B., Tao, D.: Hi-SAM: marrying segment anything model for hierarchical text segmentation (2024). arXiv preprint arXiv:2401.17904","DOI":"10.1109\/TPAMI.2024.3495831"},{"key":"3809_CR30","doi-asserted-by":"crossref","unstructured":"Shi, Z., Sun, Y., Zhang, M.: Training-free object counting with prompts. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 323\u2013331 (2024)","DOI":"10.1109\/WACV57701.2024.00039"},{"key":"3809_CR31","doi-asserted-by":"crossref","unstructured":"Wang, X., Wang, W., Cao, Y., Shen, C., Huang, T.: Images speak in images: a generalist painter for in-context visual learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6830\u20136839 (2023)","DOI":"10.1109\/CVPR52729.2023.00660"},{"key":"3809_CR32","doi-asserted-by":"crossref","unstructured":"Wang, X., Zhang, X., Cao, Y., Wang, W., Shen, C., Huang, T.: SegGPT: segmenting everything in context (2023). arXiv preprint arXiv:2304.03284","DOI":"10.1109\/ICCV51070.2023.00110"},{"key":"3809_CR33","unstructured":"Liu, Y., Zhu, M., Li, H., Chen, H., Wang, X., Shen, C.: Matcher: segment anything with one shot using all-purpose feature matching (2023). arXiv preprint arXiv:2305.13310"},{"key":"3809_CR34","unstructured":"Zhang, R., Jiang, Z., Guo, Z., Yan, S., Pan, J., Ma, X., Dong, H., Gao, P., Li, H.: Personalize segment anything model with one shot. arXiv preprint arXiv:2305.03048 (2023)"},{"key":"3809_CR35","volume-title":"Moment Functions in Image Analysis: Theory and Applications","author":"R Mukundan","year":"1998","unstructured":"Mukundan, R., Ramakrishnan, K.: Moment Functions in Image Analysis: Theory and Applications. World scientific, Singapore (1998)"},{"issue":"11","key":"3809_CR36","first-page":"120","volume":"25","author":"G Bradski","year":"2000","unstructured":"Bradski, G.: The OpenCV library. Dr. Dobb\u2019s J.: Softw. Tools Prof. Program. 25(11), 120\u2013123 (2000)","journal-title":"Dr. Dobb\u2019s J.: Softw. Tools Prof. Program."},{"key":"3809_CR37","unstructured":"Zhang, X., Wei, Y., Yang, Y., Huang, T.S.: SG-One: similarity guidance network for one-shot semantic segmentation (2018). arXiv preprint arXiv:1810.09091"},{"key":"3809_CR38","doi-asserted-by":"crossref","unstructured":"Wang, K., Liew, J.H., Zou, Y., Zhou, D., Feng, J.: Panet: Few-shot image semantic segmentation with prototype alignment. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9197\u20139206 (2019)","DOI":"10.1109\/ICCV.2019.00929"},{"key":"3809_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TPAMI.2020.3032166","volume":"01","author":"Z Tian","year":"2020","unstructured":"Tian, Z., Zhao, H., Shu, M., Yang, Z., Li, R., Jia, J.: Prior guided feature enrichment network for few-shot segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 01, 1\u20131 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3809_CR40","doi-asserted-by":"crossref","unstructured":"Lang, C., Cheng, G., Tu, B., Han, J.: Learning what not to segment: a new perspective on few-shot segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8057\u20138067 (2022)","DOI":"10.1109\/CVPR52688.2022.00789"},{"key":"3809_CR41","doi-asserted-by":"crossref","unstructured":"Zhang, C., Lin, G., Liu, F., Yao, R., Shen, C.: CANet: class-agnostic segmentation networks with iterative refinement and attentive few-shot learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5217\u20135226 (2019)","DOI":"10.1109\/CVPR.2019.00536"},{"key":"3809_CR42","doi-asserted-by":"crossref","unstructured":"Zhang, C., Lin, G., Liu, F., Guo, J., Wu, Q., Yao, R.: Pyramid graph networks with connection attentions for region-based one-shot semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9587\u20139595 (2019)","DOI":"10.1109\/ICCV.2019.00968"},{"key":"3809_CR43","first-page":"21984","volume":"34","author":"G Zhang","year":"2021","unstructured":"Zhang, G., Kang, G., Yang, Y., Wei, Y.: Few-shot segmentation via cycle-consistent transformer. Adv. Neural. Inf. Process. Syst. 34, 21984\u201321996 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3809_CR44","doi-asserted-by":"crossref","unstructured":"Lu, Z., He, S., Zhu, X., Zhang, L., Song, Y.Z., Xiang, T.: Simpler is better: few-shot semantic segmentation with classifier weight transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8741\u20138750 (2021)","DOI":"10.1109\/ICCV48922.2021.00862"},{"key":"3809_CR45","doi-asserted-by":"crossref","unstructured":"Lang, C., Tu, B., Cheng, G., Han, J.: Beyond the prototype: divide-and-conquer proxies for few-shot segmentation. arXiv preprint arXiv:2204.09903 (2022)","DOI":"10.24963\/ijcai.2022\/143"},{"key":"3809_CR46","doi-asserted-by":"crossref","unstructured":"Shi, X., Wei, D., Zhang, Y., Lu, D., Ning, M., Chen, J., Ma, K., Zheng, Y.: Dense cross-query-and-support attention weighted mask aggregation for few-shot segmentation. In: European Conference on Computer Vision, pp. 151\u2013168. Springer, Berlin (2022)","DOI":"10.1007\/978-3-031-20044-1_9"},{"key":"3809_CR47","doi-asserted-by":"crossref","unstructured":"Wang, Y., Sun, R., Zhang, T.: Rethinking the correlation in few-shot segmentation: a buoys view. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7183\u20137192 (2023)","DOI":"10.1109\/CVPR52729.2023.00694"},{"key":"3809_CR48","doi-asserted-by":"crossref","unstructured":"Nguyen, K., Todorovic, S.: Feature weighting and boosting for few-shot segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 622\u2013631 (2019)","DOI":"10.1109\/ICCV.2019.00071"},{"key":"3809_CR49","doi-asserted-by":"crossref","unstructured":"Jha, D., Smedsrud, P.H., Riegler, M.A., Halvorsen, P., De\u00a0Lange, T., Johansen, D., Johansen, H.D.: Kvasir-SEG: a segmented polyp dataset. In: MultiMedia Modeling: 26th International Conference, MMM 2020, Daejeon, South Korea, January 5\u20138, 2020, proceedings, part II 26, pp. 451\u2013462. Springer, Berlin (2020)","DOI":"10.1007\/978-3-030-37734-2_37"},{"issue":"2","key":"3809_CR50","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vision 88(2), 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vision"},{"key":"3809_CR51","doi-asserted-by":"crossref","unstructured":"Hariharan, B., Arbel\u00e1ez, P., Girshick, R., Malik, J.: Simultaneous detection and segmentation. In: Computer Vision\u2014ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part VII 13, pp. 297\u2013312. Springer, Berlin (2014)","DOI":"10.1007\/978-3-319-10584-0_20"},{"key":"3809_CR52","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft COCO: common objects in context. In: European Conference on Computer Vision. Springer, Berlin, pp. 740\u2013755 (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"3809_CR53","doi-asserted-by":"crossref","unstructured":"Liu, W., Zhang, C., Lin, G., Liu, F.: CRNet: cross-reference networks for few-shot segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2020), pp. 4165\u20134173","DOI":"10.1109\/CVPR42600.2020.00422"},{"key":"3809_CR54","unstructured":"Liu, J., Qin, Y.: Prototype refinement network for few-shot segmentation (2020). arXiv preprint arXiv:2002.03579"},{"key":"3809_CR55","doi-asserted-by":"crossref","unstructured":"Liu, Y., Zhang, X., Zhang, S., He, X.: Part-aware prototype network for few-shot semantic segmentation. In: European Conference on Computer Vision, pp. 142\u2013158. Springer, Berlin (2020)","DOI":"10.1007\/978-3-030-58545-7_9"},{"key":"3809_CR56","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"3809_CR57","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3809_CR58","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. 115, 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vis."},{"key":"3809_CR59","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234\u2013241. Springer, Berlin (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"3809_CR60","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.isprsjprs.2020.01.013","volume":"162","author":"FI Diakogiannis","year":"2020","unstructured":"Diakogiannis, F.I., Waldner, F., Caccetta, P., Wu, C.: ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogram. Remote. Sens. 162, 94\u2013114 (2020)","journal-title":"ISPRS J. Photogram. Remote. Sens."},{"key":"3809_CR61","doi-asserted-by":"crossref","unstructured":"Jha, D., Smedsrud, P.H., Riegler, M.A., Johansen, D., De\u00a0Lange, T., Halvorsen, P., Johansen, H.D.: Resunet++: an advanced architecture for medical image segmentation. In: 2019 IEEE International Symposium on Multimedia (ISM), IEEE, pp. 225\u20132255 (2019)","DOI":"10.1109\/ISM46123.2019.00049"},{"key":"3809_CR62","doi-asserted-by":"crossref","unstructured":"Wang, H., Yang, Y., Cao, X., Zhen, X., Snoek, C., Shao, L.: Variational prototype inference for few-shot semantic segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 525\u2013534 (2021)","DOI":"10.1109\/WACV48630.2021.00057"},{"issue":"3","key":"3809_CR63","doi-asserted-by":"publisher","first-page":"828","DOI":"10.1002\/ima.22428","volume":"30","author":"H Zhao","year":"2020","unstructured":"Zhao, H., Qiu, X., Lu, W., Huang, H., Jin, X.: High-quality retinal vessel segmentation using generative adversarial network with a large receptive field. Int. J. Imaging Syst. Technol. 30(3), 828\u2013842 (2020)","journal-title":"Int. J. Imaging Syst. Technol."},{"key":"3809_CR64","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.107454","volume":"128","author":"Q Qin","year":"2024","unstructured":"Qin, Q., Chen, Y.: A review of retinal vessel segmentation for fundus image analysis. Eng. Appl. Artif. Intell. 128, 107454 (2024)","journal-title":"Eng. Appl. Artif. Intell."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-03809-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-03809-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-03809-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T06:37:37Z","timestamp":1757140657000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-03809-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,26]]},"references-count":64,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["3809"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-03809-9","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,26]]},"assertion":[{"value":"9 January 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 February 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}