{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T03:29:45Z","timestamp":1775618985008,"version":"3.50.1"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"13","license":[{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"The Research Project of Key Laboratory of Fujian Province University, China","award":["ZZKY202204"],"award-info":[{"award-number":["ZZKY202204"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s10489-025-06780-z","type":"journal-article","created":{"date-parts":[[2025,8,8]],"date-time":"2025-08-08T01:28:34Z","timestamp":1754616514000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["SSRepVM-UNet: a lightweight hybrid model for medical image segmentation based on channel parallelism"],"prefix":"10.1007","volume":"55","author":[{"given":"Yijing","family":"Guo","sequence":"first","affiliation":[]},{"given":"Fuhang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Kunhua","family":"Li","sequence":"additional","affiliation":[]},{"given":"Huawei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Pengyu","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,8]]},"reference":[{"issue":"1","key":"6780_CR1","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1038\/s41598-020-57866-2","volume":"10","author":"RN D\u2019souza","year":"2020","unstructured":"D\u2019souza RN, Huang PY, Yeh FC (2020) Structural analysis and optimization of convolutional neural networks with a small sample size. Sci Rep 10(1):834","journal-title":"Sci Rep"},{"key":"6780_CR2","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: International conference on learning representations (ICLR)"},{"key":"6780_CR3","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"6780_CR4","doi-asserted-by":"crossref","unstructured":"Huang G, Liu Z, Van Der Maaten L et al (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700\u20134708","DOI":"10.1109\/CVPR.2017.243"},{"key":"6780_CR5","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu CY et al (2022) A convnet for the 2020s. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11,976\u201311,986","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"6780_CR6","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A et al (2021) An image is worth 16x16 words: Transformers for image recognition at scale. In: International conference on learning representations (ICLR)"},{"key":"6780_CR7","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y et al (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 10,012\u201310,022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"6780_CR8","unstructured":"Touvron H, Cord M, Douze M et al (2021) Training data-efficient image transformers & distillation through attention. In: International conference on machine learning, PMLR, pp 10,347\u201310,357"},{"key":"6780_CR9","unstructured":"Zhang X, Tian Y, Xie L et al (2023) Hivit: A simpler and more efficient design of hierarchical vision transformer. In: The eleventh international conference on learning representations"},{"key":"6780_CR10","doi-asserted-by":"crossref","unstructured":"Ding X, Zhang X, Han J, et al (2022) Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11,963\u201311,975","DOI":"10.1109\/CVPR52688.2022.01166"},{"key":"6780_CR11","unstructured":"Gu A, Dao T (2023) Mamba: Linear-time sequence modeling with selective state spaces. arXiv:2312.00752"},{"key":"6780_CR12","doi-asserted-by":"crossref","unstructured":"Wu R, Liu Y, Liang P, et al (2025) H-vmunet: High-order vision mamba unet for medical image segmentation. Neurocomputing p 129447","DOI":"10.1016\/j.neucom.2025.129447"},{"key":"6780_CR13","doi-asserted-by":"crossref","unstructured":"Wu R, Liu Y, Liang P et al (2024) Ultralight vm-unet: Parallel vision mamba significantly reduces parameters for skin lesion segmentation. arXiv preprint arXiv:2403.20035","DOI":"10.1016\/j.patter.2025.101298"},{"key":"6780_CR14","unstructured":"Liu Y, Tian Y, Zhao Y et al (2024) Vmamba: Visual state space model. In: Globerson A, Mackey L, Belgrave D et al (eds) Advances in neural information processing systems, vol 37. Curran Associates, Inc., pp 103,031\u2013103,063"},{"key":"6780_CR15","doi-asserted-by":"crossref","unstructured":"Ruan J, Xiang S (2024) Vm-unet: Vision mamba unet for medical image segmentation. arXiv:2402.02491","DOI":"10.1145\/3767748"},{"key":"6780_CR16","unstructured":"Yue Y, Li Z (2024) Medmamba: Vision mamba for medical image classification. arXiv preprint arXiv:2403.03849"},{"key":"6780_CR17","unstructured":"Wang A, Chen H, Lin Z et al (2024) Repvit: Revisiting mobile cnn from vit perspective. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern Recognition, pp 15,909\u201315,920"},{"issue":"104","key":"6780_CR18","first-page":"398","volume":"80","author":"S Maurya","year":"2023","unstructured":"Maurya S, Tiwari S, Mothukuri MC et al (2023) A review on recent developments in cancer detection using machine learning and deep learning models. Biomed Signal Process Control 80(104):398","journal-title":"Biomed Signal Process Control"},{"issue":"3","key":"6780_CR19","doi-asserted-by":"publisher","first-page":"1224","DOI":"10.3390\/su13031224","volume":"13","author":"X Liu","year":"2021","unstructured":"Liu X, Song L, Liu S et al (2021) A review of deep-learning-based medical image segmentation methods. Sustainability 13(3):1224","journal-title":"Sustainability"},{"key":"6780_CR20","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"6780_CR21","doi-asserted-by":"crossref","unstructured":"Peng Y, Chen DZ, Sonka M (2025) U-net v2: Rethinking the skip connections of u-net for medical image segmentation. In: 2025 IEEE 22nd International symposium on biomedical imaging (ISBI), IEEE, pp 1\u20135","DOI":"10.1109\/ISBI60581.2025.10980742"},{"key":"6780_CR22","unstructured":"Oktay O, Schlemper J, Le Folgoc L et al (2018) Attention u-net: Learning where to look for the pancreas. In: Medical imaging with deep learning, (MIDL) 2018, international conference on"},{"key":"6780_CR23","doi-asserted-by":"crossref","unstructured":"Aghdam EK, Azad R, Zarvani M et al (2023) Attention swin u-net: Cross-contextual attention mechanism for skin lesion segmentation. In: 2023 IEEE 20th International symposium on biomedical imaging (ISBI), IEEE, pp 1\u20135","DOI":"10.1109\/ISBI53787.2023.10230337"},{"key":"6780_CR24","doi-asserted-by":"crossref","unstructured":"Cao H, Wang Y, Chen J et al (2022) Swin-unet: Unet-like pure transformer for medical image segmentation. In: European conference on computer vision, Springer, pp 205\u2013218","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"6780_CR25","unstructured":"Chen J, Lu Y, Yu Q et al (2021) Transunet: Transformers make strong encoders for medical image segmentation. In: Proceedings of the AAAI conference on artificial intelligence, pp 1\u20139"},{"issue":"9","key":"6780_CR26","doi-asserted-by":"publisher","first-page":"2763","DOI":"10.1109\/TMI.2023.3264513","volume":"42","author":"A He","year":"2023","unstructured":"He A, Wang K, Li T et al (2023) H2former: An efficient hierarchical hybrid transformer for medical image segmentation. IEEE Trans Med Imaging 42(9):2763\u20132775","journal-title":"IEEE Trans Med Imaging"},{"key":"6780_CR27","unstructured":"Ma J, Li F, Wang B (2024) U-mamba: Enhancing long-range dependency for biomedical image segmentation. arXiv preprint arXiv:2401.04722"},{"key":"6780_CR28","doi-asserted-by":"crossref","unstructured":"Liu J, Yang H, Zhou HY et al (2024) Swin-umamba: Mamba-based unet with imagenet-based pretraining. In: International conference on medical image computing and computer-assisted intervention, Springer, pp 615\u2013625","DOI":"10.1007\/978-3-031-72114-4_59"},{"key":"6780_CR29","unstructured":"Xu J (2024) Hc-mamba: Vision mamba with hybrid convolutional techniques for medical image segmentation. arXiv:2405.05007"},{"key":"6780_CR30","doi-asserted-by":"crossref","unstructured":"Yu W, Luo M, Zhou P et al (2022) Metaformer is actually what you need for vision. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10,819\u201310,829","DOI":"10.1109\/CVPR52688.2022.01055"},{"issue":"125","key":"6780_CR31","first-page":"518","volume":"261","author":"G Li","year":"2025","unstructured":"Li G, Huang Q, Wang W et al (2025) Selective and multi-scale fusion mamba for medical image segmentation. Expert Syst Appl 261(125):518","journal-title":"Expert Syst Appl"},{"key":"6780_CR32","doi-asserted-by":"crossref","unstructured":"Ruan J, Xiang S, Xie M et al (2022) Malunet: A multi-attention and light-weight unet for skin lesion segmentation. In: 2022 IEEE International conference on bioinformatics and biomedicine (BIBM), IEEE, pp 1150\u20131156","DOI":"10.1109\/BIBM55620.2022.9995040"},{"key":"6780_CR33","unstructured":"Gutman D, Codella NC, Celebi E et al (2016) Skin lesion analysis toward melanoma detection: A challenge at the international symposium on biomedical imaging (isbi) 2016, hosted by the international skin imaging collaboration (isic). arXiv:1605.01397"},{"key":"6780_CR34","doi-asserted-by":"crossref","unstructured":"Codella NC, Gutman D, Celebi ME et al (2018) Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collaboration (isic). In: 2018 IEEE 15th International symposium on biomedical imaging (ISBI 2018). IEEE, pp 168\u2013172","DOI":"10.1109\/ISBI.2018.8363547"},{"key":"6780_CR35","unstructured":"Codella N, Rotemberg V, Tschandl P et al (2019) Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic). arXiv:1902.03368"},{"key":"6780_CR36","doi-asserted-by":"crossref","unstructured":"Esteva A, Kuprel B, Novoa RA et al (2017) Dermatologist-level classification of skin cancer with deep neural networks. nature 542(7639):115\u2013118","DOI":"10.1038\/nature21056"},{"key":"6780_CR37","doi-asserted-by":"crossref","unstructured":"Jha D, Smedsrud PH, Riegler MA et al (2020) Kvasir-seg: A segmented polyp dataset. In: MultiMedia modeling: 26th international conference, MMM 2020, Daejeon, South Korea, January 5\u20138, 2020, proceedings, part II 26, Springer, pp 451\u2013462","DOI":"10.1007\/978-3-030-37734-2_37"},{"key":"6780_CR38","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.compmedimag.2015.02.007","volume":"43","author":"J Bernal","year":"2015","unstructured":"Bernal J, S\u00e1nchez FJ, Fern\u00e1ndez-Esparrach G et al (2015) Wm-dova maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Comput Med Imaging Graph 43:99\u2013111","journal-title":"Comput Med Imaging Graph"},{"key":"6780_CR39","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: Medical image computing and computer-assisted intervention\u2013MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18, Springer, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"6780_CR40","doi-asserted-by":"crossref","unstructured":"Wu H, Zhong J, Wang W et al (2021) Precise yet efficient semantic calibration and refinement in convnets for real-time polyp segmentation from colonoscopy videos. In: Proceedings of the AAAI conference on artificial intelligence, pp 2916\u20132924","DOI":"10.1609\/aaai.v35i4.16398"},{"key":"6780_CR41","doi-asserted-by":"crossref","unstructured":"Hu S, Liao Z, Xia Y (2023) Devil is in channels: Contrastive single domain generalization for medical image segmentation. In: International conference on medical image computing and computer-assisted intervention, Springer, pp 14\u201323","DOI":"10.1007\/978-3-031-43901-8_2"},{"key":"6780_CR42","doi-asserted-by":"crossref","unstructured":"Zhang M, Yu Y, Jin S et al (2024) Vm-unet-v2: rethinking vision mamba unet for medical image segmentation. In: International symposium on bioinformatics research and applications, Springer, pp 335\u2013346","DOI":"10.1007\/978-981-97-5128-0_27"},{"key":"6780_CR43","unstructured":"Liao W, Zhu Y, Wang X et al (2024) Lightm-unet: Mamba assists in lightweight unet for medical image segmentation. arXiv:2403.05246"},{"key":"6780_CR44","doi-asserted-by":"crossref","unstructured":"Zhang Y, Liu H, Hu Q (2021) Transfuse: Fusing transformers and cnns for medical image segmentation. In: Medical image computing and computer assisted intervention\u2013MICCAI 2021: 24th international conference, Strasbourg, France, September 27\u2013October 1, 2021, proceedings, Part I 24, Springer, pp 14\u201324","DOI":"10.1007\/978-3-030-87193-2_2"},{"issue":"2","key":"6780_CR45","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1109\/JBHI.2018.2859898","volume":"23","author":"H Li","year":"2018","unstructured":"Li H, He X, Zhou F et al (2018) Dense deconvolutional network for skin lesion segmentation. IEEE J Biomed Health Inform 23(2):527\u2013537","journal-title":"IEEE J Biomed Health Inform"},{"issue":"101","key":"6780_CR46","first-page":"658","volume":"78","author":"E Nasr-Esfahani","year":"2019","unstructured":"Nasr-Esfahani E, Rafiei S, Jafari MH et al (2019) Dense pooling layers in fully convolutional network for skin lesion segmentation. Comput Med Imaging Graph 78(101):658","journal-title":"Comput Med Imaging Graph"},{"key":"6780_CR47","doi-asserted-by":"publisher","first-page":"289","DOI":"10.1016\/j.cmpb.2019.07.005","volume":"178","author":"P Tang","year":"2019","unstructured":"Tang P, Liang Q, Yan X et al (2019) Efficient skin lesion segmentation using separable-unet with stochastic weight averaging. Comput Methods Programs Biomed 178:289\u2013301","journal-title":"Comput Methods Programs Biomed"},{"key":"6780_CR48","doi-asserted-by":"crossref","unstructured":"Lee HJ, Kim JU, Lee S et al (2020) Structure boundary preserving segmentation for medical image with ambiguous boundary. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4817\u20134826","DOI":"10.1109\/CVPR42600.2020.00487"},{"issue":"10","key":"6780_CR49","doi-asserted-by":"publisher","first-page":"3008","DOI":"10.1109\/TMI.2020.2983721","volume":"39","author":"S Feng","year":"2020","unstructured":"Feng S, Zhao H, Shi F et al (2020) Cpfnet: Context pyramid fusion network for medical image segmentation. IEEE Trans Med Imaging 39(10):3008\u20133018","journal-title":"IEEE Trans Med Imaging"},{"issue":"101","key":"6780_CR50","first-page":"716","volume":"64","author":"B Lei","year":"2020","unstructured":"Lei B, Xia Z, Jiang F et al (2020) Skin lesion segmentation via generative adversarial networks with dual discriminators. Med Image Anal 64(101):716","journal-title":"Med Image Anal"},{"issue":"102","key":"6780_CR51","first-page":"327","volume":"76","author":"H Wu","year":"2022","unstructured":"Wu H, Chen S, Chen G et al (2022) Fat-net: Feature adaptive transformers for automated skin lesion segmentation. Med Image Anal 76(102):327","journal-title":"Med Image Anal"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-025-06780-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-025-06780-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-025-06780-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,27]],"date-time":"2025-09-27T14:34:36Z","timestamp":1758983676000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-025-06780-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8]]},"references-count":51,"journal-issue":{"issue":"13","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["6780"],"URL":"https:\/\/doi.org\/10.1007\/s10489-025-06780-z","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8]]},"assertion":[{"value":"12 July 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 August 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All the authors declare that they have no competing financial interests or personal relationships that could influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"This article does not contain studies with human participants or animals. Statement of informed consent is not applicable since the manuscript does not contain any patient data.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}}],"article-number":"911"}}