{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T10:36:53Z","timestamp":1773225413543,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T00:00:00Z","timestamp":1773100800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T00:00:00Z","timestamp":1773100800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62402202"],"award-info":[{"award-number":["62402202"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2024QF036"],"award-info":[{"award-number":["ZR2024QF036"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shandong Province\u2019s Undergraduate Education Reform Project","award":["M2023218"],"award-info":[{"award-number":["M2023218"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1007\/s00530-026-02285-y","type":"journal-article","created":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T14:04:08Z","timestamp":1773151448000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Resukan: an improved medical image segmentation method based on U-KAN architecture"],"prefix":"10.1007","volume":"32","author":[{"given":"Yu","family":"Feng","sequence":"first","affiliation":[]},{"given":"Haomeng","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Chuanshen","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Haiyong","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,10]]},"reference":[{"key":"2285_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2016.06.023","author":"J Weese","year":"2016","unstructured":"Weese, J., Lorenz, C.: Four challenges in medical image analysis from an industrial perspective. Elsevier (2016). https:\/\/doi.org\/10.1016\/j.media.2016.06.023","journal-title":"Elsevier"},{"key":"2285_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102385","volume":"78","author":"S Hansen","year":"2022","unstructured":"Hansen, S., Gautam, S., Jenssen, R., Kampffmeyer, M.: Anomaly detection-inspired few-shot medical image segmentation through self-supervision with supervoxels. Med. Image Anal. 78, 102385 (2022). https:\/\/doi.org\/10.1016\/j.media.2022.102385","journal-title":"Med. Image Anal."},{"issue":"1","key":"2285_CR3","doi-asserted-by":"publisher","first-page":"654","DOI":"10.48550\/arXiv.2304.12306","volume":"15","author":"J Ma","year":"2024","unstructured":"Ma, J., He, Y., Li, F., Han, L., You, C., Wang, B.: Segment anything in medical images. Nat. Commun. 15(1), 654 (2024). https:\/\/doi.org\/10.48550\/arXiv.2304.12306","journal-title":"Nat. Commun."},{"key":"2285_CR4","first-page":"696","volume-title":"Neural networks: a comprehensive foundation","author":"S Haykin","year":"1994","unstructured":"Haykin, S.: Neural networks: a comprehensive foundation, p. 696. Prentice Hall PTR, New York, USA (1994)"},{"issue":"4","key":"2285_CR5","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/BF02551274","volume":"2","author":"G Cybenko","year":"1989","unstructured":"Cybenko, G.: Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. 2(4), 303\u2013314 (1989). https:\/\/doi.org\/10.1007\/BF02551274","journal-title":"Math. Control Signals Syst."},{"issue":"5","key":"2285_CR6","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1016\/0893-6080(89)90020-8","volume":"2","author":"K Hornik","year":"1989","unstructured":"Hornik, K., Stinchcombe, M., White, H.: Multilayer feedforward networks are universal approximators. Neural Netw. 2(5), 359\u2013366 (1989). https:\/\/doi.org\/10.1016\/0893-6080(89)90020-8","journal-title":"Neural Netw."},{"issue":"4","key":"2285_CR7","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1007\/s11760-022-02325-w","volume":"17","author":"I Bakkouri","year":"2023","unstructured":"Bakkouri, I., Afdel, K.: Mlca2f: multi-level context attentional feature fusion for covid-19 lesion segmentation from ct scans. SIViP 17(4), 1181\u20131188 (2023). https:\/\/doi.org\/10.1007\/s11760-022-02325-w","journal-title":"SIViP"},{"issue":"10","key":"2285_CR8","doi-asserted-by":"publisher","first-page":"12939","DOI":"10.1007\/s11042-018-6267-z","volume":"78","author":"I Bakkouri","year":"2019","unstructured":"Bakkouri, I., Afdel, K.: Multi-scale cnn based on region proposals for efficient breast abnormality recognition. Multimedia Tools Appl 78(10), 12939\u201312960 (2019). https:\/\/doi.org\/10.1007\/s11042-018-6267-z","journal-title":"Multimedia Tools Appl"},{"issue":"29","key":"2285_CR9","doi-asserted-by":"publisher","first-page":"20483","DOI":"10.1007\/s11042-019-07988-1","volume":"79","author":"I Bakkouri","year":"2020","unstructured":"Bakkouri, I., Afdel, K.: Computer-aided diagnosis (cad) system based on multi-layer feature fusion network for skin lesion recognition in dermoscopy images. Multimedia Tools Appl 79(29), 20483\u201320518 (2020). https:\/\/doi.org\/10.1007\/s11042-019-07988-1","journal-title":"Multimedia Tools Appl"},{"key":"2285_CR10","doi-asserted-by":"publisher","unstructured":"Bakkouri, I., Afdel, K., Benois-Pineau, J., Initiative, G.C.F.t.A.D.N.: Bg-3dm2f: bidirectional gated 3d multi-scale feature fusion for alzheimer\u2019s disease diagnosis. Multimedia Tools and Applications 81(8), 10743\u201310776 (2022) https:\/\/doi.org\/10.1007\/s11042-022-12242-2","DOI":"10.1007\/s11042-022-12242-2"},{"key":"2285_CR11","doi-asserted-by":"publisher","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Advances in neural information processing systems 30 (2017) https:\/\/doi.org\/10.48550\/arXiv.1706.03762","DOI":"10.48550\/arXiv.1706.03762"},{"key":"2285_CR12","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1162\/tacl_a_00306","volume":"8","author":"M Hahn","year":"2020","unstructured":"Hahn, M.: Theoretical limitations of self-attention in neural sequence models. Trans Assoc Comput Linguistics 8, 156\u2013171 (2020). https:\/\/doi.org\/10.1162\/tacl_a_00306","journal-title":"Trans Assoc Comput Linguistics"},{"key":"2285_CR13","doi-asserted-by":"publisher","unstructured":"Han, S., Mao, H., Dally, W.J.: Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015) https:\/\/doi.org\/10.48550\/arXiv.1510.00149","DOI":"10.48550\/arXiv.1510.00149"},{"key":"2285_CR14","doi-asserted-by":"publisher","unstructured":"Chen, T., Xu, B., Zhang, C., Guestrin, C.: Training deep nets with sublinear memory cost. arXiv preprint arXiv:1604.06174 (2016) https:\/\/doi.org\/10.48550\/arXiv.1604.06174","DOI":"10.48550\/arXiv.1604.06174"},{"key":"2285_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.dsp.2017.10.011","volume":"73","author":"G Montavon","year":"2018","unstructured":"Montavon, G., Samek, W., M\u00fcller, K.-R.: Methods for interpreting and understanding deep neural networks. Digital signal processing 73, 1\u201315 (2018). https:\/\/doi.org\/10.1016\/j.dsp.2017.10.011","journal-title":"Digital signal processing"},{"key":"2285_CR16","doi-asserted-by":"publisher","unstructured":"Cunningham, H., Ewart, A., Riggs, L., Huben, R., Sharkey, L.: Sparse autoencoders find highly interpretable features in language models. arXiv preprint arXiv:2309.08600 (2023) https:\/\/doi.org\/10.48550\/arXiv.2309.08600","DOI":"10.48550\/arXiv.2309.08600"},{"issue":"5","key":"2285_CR17","doi-asserted-by":"publisher","first-page":"206","DOI":"10.48550\/arXiv.1811.10154","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Mach Intell 1(5), 206\u2013215 (2019). https:\/\/doi.org\/10.48550\/arXiv.1811.10154","journal-title":"Nature Mach Intell"},{"key":"2285_CR18","doi-asserted-by":"publisher","unstructured":"Liu, Z., Wang, Y., Vaidya, S., Ruehle, F., Halverson, J., Solja\u010di\u0107, M., Hou, T.Y., Tegmark, M.: Kan: Kolmogorov-arnold networks. arXiv preprint arXiv:2404.19756 (2024) https:\/\/doi.org\/10.48550\/arXiv.2404.19756","DOI":"10.48550\/arXiv.2404.19756"},{"key":"2285_CR19","doi-asserted-by":"crossref","unstructured":"Kolmogorov, A.N.: On the Representation of Continuous Functions of Several Variables by Superpositions of Continuous Functions of a Smaller Number of Variables, vol. 17, pp. 369\u2013373. American Mathematical Society, Providence, RI, USA (1961)","DOI":"10.1090\/trans2\/017\/12"},{"key":"2285_CR20","unstructured":"Kolmogorov, A.N.: On the representations of continuous functions of many variables by superposition of continuous functions of one variable and addition. In: Dokl. Akad. Nauk USSR, vol. 114, pp. 953\u2013956 (1957)"},{"key":"2285_CR21","doi-asserted-by":"publisher","first-page":"653","DOI":"10.1007\/s00365-009-9054-2","volume":"30","author":"J Braun","year":"2009","unstructured":"Braun, J., Griebel, M.: On a constructive proof of kolmogorov\u2019s superposition theorem. Constr. Approx. 30, 653\u2013675 (2009). https:\/\/doi.org\/10.1007\/s00365-009-9054-2","journal-title":"Constr. Approx."},{"issue":"1","key":"2285_CR22","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1162\/neco.1993.5.1.18","volume":"5","author":"J-N Lin","year":"1993","unstructured":"Lin, J.-N., Unbehauen, R.: On the realization of a kolmogorov network. Neural Comput. 5(1), 18\u201320 (1993). https:\/\/doi.org\/10.1162\/neco.1993.5.1.18","journal-title":"Neural Comput."},{"issue":"1","key":"2285_CR23","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1016\/s0893-6080(01)00107-1","volume":"15","author":"DA Sprecher","year":"2002","unstructured":"Sprecher, D.A., Draghici, S.: Space-filling curves and kolmogorov superposition-based neural networks. Neural Netw. 15(1), 57\u201367 (2002). https:\/\/doi.org\/10.1016\/s0893-6080(01)00107-1","journal-title":"Neural Netw."},{"key":"2285_CR24","doi-asserted-by":"publisher","unstructured":"K\u00f6ppen, M.: On the training of a kolmogorov network. In: Artificial Neural Networks\u2014ICANN 2002: International Conference Madrid, Spain, August 28\u201330, 2002 Proceedings 12, pp. 474\u2013479 (2002). https:\/\/doi.org\/10.1007\/3-540-46084-5_77","DOI":"10.1007\/3-540-46084-5_77"},{"key":"2285_CR25","doi-asserted-by":"publisher","unstructured":"Leni, P.-E., Fougerolle, Y.D., Truchetet, F.: The kolmogorov spline network for image processing. In: Image processing: concepts, methodologies, tools, and applications, pp. 54\u201378 (2013). https:\/\/doi.org\/10.4018\/978-1-4666-3994-2.ch004","DOI":"10.4018\/978-1-4666-3994-2.ch004"},{"key":"2285_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.neunet.2019.12.013","volume":"129","author":"H Montanelli","year":"2020","unstructured":"Montanelli, H., Yang, H.: Error bounds for deep relu networks using the kolmogorov-arnold superposition theorem. Neural Netw. 129, 1\u20136 (2020). https:\/\/doi.org\/10.1016\/j.neunet.2019.12.013","journal-title":"Neural Netw."},{"key":"2285_CR27","doi-asserted-by":"publisher","unstructured":"Lai, M.-J., Shen, Z.: The kolmogorov superposition theorem can break the curse of dimensionality when approximating high dimensional functions. arXiv preprint arXiv:2112.09963 (2021) https:\/\/doi.org\/10.48550\/arXiv.2112.09963","DOI":"10.48550\/arXiv.2112.09963"},{"key":"2285_CR28","doi-asserted-by":"publisher","first-page":"332","DOI":"10.1016\/j.neunet.2022.04.029","volume":"152","author":"D Fakhoury","year":"2022","unstructured":"Fakhoury, D., Fakhoury, E., Speleers, H.: Exsplinet: an interpretable and expressive spline-based neural network. Neural Netw. 152, 332\u2013346 (2022). https:\/\/doi.org\/10.1016\/j.neunet.2022.04.029","journal-title":"Neural Netw."},{"key":"2285_CR29","doi-asserted-by":"publisher","unstructured":"He, J.: On the optimal expressive power of relu dnns and its application in approximation with kolmogorov superposition theorem. arXiv preprint arXiv:2308.05509 (2023) https:\/\/doi.org\/10.48550\/arXiv.2308.05509","DOI":"10.48550\/arXiv.2308.05509"},{"key":"2285_CR30","doi-asserted-by":"publisher","unstructured":"Li, C., Liu, X., Li, W., Wang, C., Liu, H., Liu, Y., Chen, Z., Yuan, Y.: U-kan makes strong backbone for medical image segmentation and generation. In: Proceedings of the AAAI conference on artificial intelligence, vol. 39 4652\u20134660 (2025). https:\/\/doi.org\/10.48550\/arXiv.2406.02918","DOI":"10.48550\/arXiv.2406.02918"},{"key":"2285_CR31","doi-asserted-by":"publisher","unstructured":"Hanin, B., Sellke, M.: Approximating continuous functions by relu nets of minimal width. arXiv preprint arXiv:1710.11278 (2017) https:\/\/doi.org\/10.48550\/arXiv.1710.11278","DOI":"10.48550\/arXiv.1710.11278"},{"issue":"5","key":"2285_CR32","doi-asserted-by":"publisher","first-page":"503","DOI":"10.48550\/arXiv.1611.00740","volume":"14","author":"T Poggio","year":"2017","unstructured":"Poggio, T., Mhaskar, H., Rosasco, L., Miranda, B., Liao, Q.: Why and when can deep-but not shallow-networks avoid the curse of dimensionality: a review. Int. J. Autom. Comput. 14(5), 503\u2013519 (2017). https:\/\/doi.org\/10.48550\/arXiv.1611.00740","journal-title":"Int. J. Autom. Comput."},{"key":"2285_CR33","doi-asserted-by":"publisher","unstructured":"Yang, Z., Zhang, J., Luo, X., Lu, Z., Shen, L.: Medkan: An advanced kolmogorov-arnold network for medical image classification. arXiv preprint arXiv:2502.18416 (2025) https:\/\/doi.org\/10.48550\/arXiv.2502.18416","DOI":"10.48550\/arXiv.2502.18416"},{"issue":"6","key":"2285_CR34","doi-asserted-by":"publisher","first-page":"4303","DOI":"10.1109\/JBHI.2025.3541982","volume":"29","author":"G Wang","year":"2025","unstructured":"Wang, G., Zhu, Q., Song, C., Wei, B., Li, S.: Medkaformer: When kolmogorov-arnold theorem meets vision transformer for medical image representation. IEEE J. Biomed. Health Inform. 29(6), 4303\u20134313 (2025). https:\/\/doi.org\/10.1109\/JBHI.2025.3541982","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"2285_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2025.3545699","volume":"74","author":"S Lin","year":"2025","unstructured":"Lin, S., Hu, R., Li, Z., Lin, Q., Zeng, K., Wu, X.: Kac-unet: A medical image segmentation with the adaptive group strategy and kolmogorov-arnold network. IEEE Trans. Instrum. Meas. 74, 1\u201313 (2025). https:\/\/doi.org\/10.1109\/TIM.2025.3545699","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2285_CR36","doi-asserted-by":"publisher","unstructured":"Lee, Y., Gong, J., Kang, J.: A unified benchmark of federated learning with kolmogorov-arnold networks for medical imaging. arXiv preprint arXiv:2504.19639 (2025) https:\/\/doi.org\/10.48550\/arXiv.2504.19639","DOI":"10.48550\/arXiv.2504.19639"},{"key":"2285_CR37","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-assisted intervention-MICCAI 2015: 18th International Conference, Munich, Germany, October 5\u20139, 2015, Proceedings, Part III 18, pp. 234\u2013241 (2015). https:\/\/doi.org\/10.48550\/arXiv.1505.04597","DOI":"10.48550\/arXiv.1505.04597"},{"key":"2285_CR38","doi-asserted-by":"publisher","first-page":"103347","DOI":"10.1016\/j.media.2024.103347","volume":"99","author":"X Du","year":"2025","unstructured":"Du, X., Xu, X., Chen, J., Zhang, X., Li, L., Liu, H., Li, S.: Um-net: Rethinking icgnet for polyp segmentation with uncertainty modeling. Med. Image Anal. 99, 103347 (2025). https:\/\/doi.org\/10.1016\/j.media.2024.103347","journal-title":"Med. Image Anal."},{"key":"2285_CR39","doi-asserted-by":"publisher","unstructured":"Milletari, F., Navab, N., Ahmadi, S.-A.: V-net: Fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV), pp. 565\u2013571 (2016). https:\/\/doi.org\/10.48550\/arXiv.1606.04797","DOI":"10.48550\/arXiv.1606.04797"},{"key":"2285_CR40","doi-asserted-by":"publisher","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N.Y., Kainz, B., et al.: Attention u-net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018) https:\/\/doi.org\/10.48550\/arXiv.1804.03999","DOI":"10.48550\/arXiv.1804.03999"},{"key":"2285_CR41","doi-asserted-by":"publisher","unstructured":"Tang, T., Chen, Y., Shu, H.: 3d u-kan implementation for multi-modal mri brain tumor segmentation (2024) https:\/\/doi.org\/10.48550\/arXiv.2408.00273","DOI":"10.48550\/arXiv.2408.00273"},{"issue":"1","key":"2285_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/sdata.2017.117","volume":"4","author":"S Bakas","year":"2017","unstructured":"Bakas, S., Akbari, H., Sotiras, A., Bilello, M., Rozycki, M., Kirby, J.S., Freymann, J.B., Farahani, K., Davatzikos, C.: Advancing the cancer genome atlas glioma mri collections with expert segmentation labels and radiomic features. Sci Data 4(1), 1\u201313 (2017). https:\/\/doi.org\/10.1038\/sdata.2017.117","journal-title":"Sci Data"},{"key":"2285_CR43","doi-asserted-by":"publisher","unstructured":"Simpson, A.L., Antonelli, M., Bakas, S., Bilello, M., Farahani, K., Van Ginneken, B., Kopp-Schneider, A., Landman, B.A., Litjens, G., Menze, B., et al.: A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv preprint arXiv:1902.09063 (2019) https:\/\/doi.org\/10.48550\/arXiv.1902.09063","DOI":"10.48550\/arXiv.1902.09063"},{"key":"2285_CR44","doi-asserted-by":"publisher","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: Unet++: A nested u-net architecture for medical image segmentation. In: Deep learning in medical image analysis and multimodal learning for clinical decision support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4, pp. 3\u201311 (2018). https:\/\/doi.org\/10.1007\/978-3-030-00889-5_1","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"2285_CR45","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016). https:\/\/doi.org\/10.48550\/arXiv.1512.03385","DOI":"10.48550\/arXiv.1512.03385"},{"key":"2285_CR46","doi-asserted-by":"publisher","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"2285_CR47","doi-asserted-by":"publisher","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale (2020) https:\/\/doi.org\/10.48550\/arXiv.2010.11929","DOI":"10.48550\/arXiv.2010.11929"},{"key":"2285_CR48","doi-asserted-by":"publisher","unstructured":"Chen, Y., Shi, H., Liu, X., Shi, T., Zhang, R., Liu, D., Xiong, Z., Wu, F.: Tokenunify: Scalable autoregressive visual pre-training with mixture token prediction (2024) https:\/\/doi.org\/10.48550\/arXiv.2405.16847","DOI":"10.48550\/arXiv.2405.16847"},{"key":"2285_CR49","doi-asserted-by":"publisher","first-page":"3726","DOI":"10.48550\/arXiv.2006.12030","volume":"31","author":"J Cao","year":"2022","unstructured":"Cao, J., Li, Y., Sun, M., Chen, Y., Lischinski, D., Cohen-Or, D., Chen, B., Tu, C.: Do-conv: Depthwise over-parameterized convolutional layer. IEEE Trans. Image Process. 31, 3726\u20133736 (2022). https:\/\/doi.org\/10.48550\/arXiv.2006.12030","journal-title":"IEEE Trans. Image Process."},{"key":"2285_CR50","doi-asserted-by":"publisher","unstructured":"Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization (2016). https:\/\/doi.org\/10.48550\/arXiv.1607.06450","DOI":"10.48550\/arXiv.1607.06450"},{"key":"2285_CR51","doi-asserted-by":"publisher","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., K\u00f6pf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Chintala, S.: Pytorch: An imperative style, high-performance deep learning library (2019) https:\/\/doi.org\/10.48550\/arXiv.1912.01703","DOI":"10.48550\/arXiv.1912.01703"},{"key":"2285_CR52","doi-asserted-by":"publisher","unstructured":"Jha, D., Smedsrud, P.H., Riegler, M.A., Halvorsen, P., De Lange, T., Johansen, D., Johansen, H.D.: Kvasir-seg: A segmented polyp dataset. In: International Conference on Multimedia Modeling, pp. 451\u2013462 (2019). https:\/\/doi.org\/10.48550\/arXiv.1911.07069","DOI":"10.48550\/arXiv.1911.07069"},{"key":"2285_CR53","doi-asserted-by":"publisher","unstructured":"Bernal, J., S\u00e1nchez, F.J., Fern\u00e1ndez-Esparrach, G., Gil, D., Rodr\u00edguez, C., Vilari\u00f1o, F.: Wm-dova maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Computerized medical imaging and graphics 43, 99\u2013111 (2015) https:\/\/doi.org\/10.1016\/j.compmedimag.2015.02.007","DOI":"10.1016\/j.compmedimag.2015.02.007"},{"issue":"2","key":"2285_CR54","doi-asserted-by":"publisher","first-page":"630","DOI":"10.1109\/TMI.2015.2487997","volume":"35","author":"N Tajbakhsh","year":"2015","unstructured":"Tajbakhsh, N., Gurudu, S.R., Liang, J.: Automated polyp detection in colonoscopy videos using shape and context information. IEEE Trans. Med. Imaging 35(2), 630\u2013644 (2015). https:\/\/doi.org\/10.1109\/TMI.2015.2487997","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"1","key":"2285_CR55","doi-asserted-by":"publisher","first-page":"4037190","DOI":"10.48550\/arXiv.1612.00799","volume":"2017","author":"D V\u00e1zquez","year":"2017","unstructured":"V\u00e1zquez, D., Bernal, J., S\u00e1nchez, F.J., Fern\u00e1ndez-Esparrach, G., L\u00f3pez, A.M., Romero, A., Drozdzal, M., Courville, A.: A benchmark for endoluminal scene segmentation of colonoscopy images. J Healthcare Eng. 2017(1), 4037190 (2017). https:\/\/doi.org\/10.48550\/arXiv.1612.00799","journal-title":"J Healthcare Eng."},{"key":"2285_CR56","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1007\/s11548-013-0926-3","volume":"9","author":"J Silva","year":"2014","unstructured":"Silva, J., Histace, A., Romain, O., Dray, X., Granado, B.: Toward embedded detection of polyps in wce images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 9, 283\u2013293 (2014). https:\/\/doi.org\/10.1007\/s11548-013-0926-3","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"2285_CR57","doi-asserted-by":"publisher","unstructured":"Jha, D., Smedsrud, P.H., Riegler, M.A., Johansen, D., De Lange, T., Halvorsen, P., Johansen, H.D.: Resunet++: An advanced architecture for medical image segmentation. In: 2019 IEEE International Symposium on Multimedia (ISM), pp. 225\u20132255 (2019). https:\/\/doi.org\/10.1109\/ISM46123.2019.00049","DOI":"10.1109\/ISM46123.2019.00049"},{"key":"2285_CR58","doi-asserted-by":"publisher","unstructured":"Margolin, R., Zelnik-Manor, L., Tal, A.: How to evaluate foreground maps? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2014). https:\/\/doi.org\/10.1109\/CVPR.2014.39","DOI":"10.1109\/CVPR.2014.39"},{"key":"2285_CR59","doi-asserted-by":"publisher","unstructured":"Jha, D., Riegler, M.A., Johansen, D., Halvorsen, P., Johansen, H.D.: Doubleu-net: A deep convolutional neural network for medical image segmentation. In: 2020 IEEE 33rd International Symposium on Computer-based Medical Systems (CBMS), pp. 558\u2013564 (2020). https:\/\/doi.org\/10.48550\/arXiv.2006.04868","DOI":"10.48550\/arXiv.2006.04868"},{"key":"2285_CR60","doi-asserted-by":"publisher","unstructured":"Huang, C.-H., Wu, H.-Y., Lin, Y.-L.: Hardnet-mseg: A simple encoder-decoder polyp segmentation neural network that achieves over 0.9 mean dice and 86 fps (2021) https:\/\/doi.org\/10.48550\/arXiv.2101.07172","DOI":"10.48550\/arXiv.2101.07172"},{"key":"2285_CR61","doi-asserted-by":"publisher","unstructured":"Fan, D.-P., Ji, G.-P., Zhou, T., Chen, G., Fu, H., Shen, J., Shao, L.: Pranet: Parallel reverse attention network for polyp segmentation. In: International Conference on Medical Image Computing and Computer-assisted Intervention, pp. 263\u2013273 (2020). https:\/\/doi.org\/10.48550\/arXiv.2006.11392","DOI":"10.48550\/arXiv.2006.11392"},{"key":"2285_CR62","doi-asserted-by":"publisher","unstructured":"Lou, A., Guan, S., Ko, H., Loew, M.H.: Caranet: context axial reverse attention network for segmentation of small medical objects. In: Medical Imaging 2022: Image Processing, vol. 12032, pp. 81\u201392 (2022). https:\/\/doi.org\/10.48550\/arXiv.2108.07368","DOI":"10.48550\/arXiv.2108.07368"},{"key":"2285_CR63","doi-asserted-by":"publisher","unstructured":"Wang, Z., Li, T., Liu, M., Jiang, J., Liu, X.: Dcatnet: polyp segmentation with deformable convolution and contextual-aware attention network. BMC Medical Imaging 25(120) (2025) https:\/\/doi.org\/10.1186\/s12880-025-01661-w","DOI":"10.1186\/s12880-025-01661-w"},{"issue":"5","key":"2285_CR64","doi-asserted-by":"publisher","first-page":"2950","DOI":"10.1002\/mp.17660","volume":"52","author":"SF Qadri","year":"2025","unstructured":"Qadri, S.F., Rong, C., Ahmad, M., Li, J., Qadri, S., Zareen, S.S., Zhuang, Z., Khan, S., Lin, H.: Chan-vese aided fuzzy c-means approach for whole breast and fibroglandular tissue segmentation: preliminary application to real-world breast mri. Med. Phys. 52(5), 2950\u20132960 (2025). https:\/\/doi.org\/10.1002\/mp.17660","journal-title":"Med. Phys."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-026-02285-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-026-02285-y","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-026-02285-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T14:04:20Z","timestamp":1773151460000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-026-02285-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,10]]},"references-count":64,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,6]]}},"alternative-id":["2285"],"URL":"https:\/\/doi.org\/10.1007\/s00530-026-02285-y","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,10]]},"assertion":[{"value":"25 August 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"We hereby formally declare that all phases of this research strictly adhere to academic ethical standards. As the research team, we have thoroughly examined all potential factors that might compromise the objectivity of this study, including but not limited to academic collaborations, financial interests, and personal relationships. After careful review, we confirm that no actual or potential Conflict of interest exist in this work.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"202"}}