{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T07:25:53Z","timestamp":1777879553846,"version":"3.51.4"},"reference-count":52,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,8,1]],"date-time":"2026-08-01T00:00:00Z","timestamp":1785542400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100003392","name":"Fujian Provincial Natural Science Foundation","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003392","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Biomedical Signal Processing and Control"],"published-print":{"date-parts":[[2026,8]]},"DOI":"10.1016\/j.bspc.2026.110338","type":"journal-article","created":{"date-parts":[[2026,4,18]],"date-time":"2026-04-18T12:45:52Z","timestamp":1776516352000},"page":"110338","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["MSFLUNet: A multi-scale feature fusion lightweight U-Net for efficient medical image segmentation"],"prefix":"10.1016","volume":"121","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5985-6001","authenticated-orcid":false,"given":"Tian-jian","family":"Luo","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-1467-0765","authenticated-orcid":false,"given":"Qinghua","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.bspc.2026.110338_b0005","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2024.108238","article-title":"Segment anything model for medical image segmentation: Current applications and future directions","volume":"171","author":"Zhang","year":"2024","journal-title":"Comput. Biol. Med."},{"key":"10.1016\/j.bspc.2026.110338_b0010","series-title":"October). Medical Image Segmentation: A Review of Modern Architectures","first-page":"691","author":"Salpea","year":"2022"},{"issue":"12","key":"10.1016\/j.bspc.2026.110338_b0015","doi-asserted-by":"crossref","first-page":"10076","DOI":"10.1109\/TPAMI.2024.3435571","article-title":"Medical image segmentation review: The success of u-net","volume":"46","author":"Azad","year":"2024","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.bspc.2026.110338_b0020","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.110491","article-title":"UCTNet: Uncertainty-guided CNN-Transformer hybrid networks for medical image segmentation","volume":"152","author":"Guo","year":"2024","journal-title":"Pattern Recogn."},{"key":"10.1016\/j.bspc.2026.110338_b0025","series-title":"October). U-Net: Convolutional Networks for Biomedical Image Segmentation","first-page":"234","author":"Ronneberger","year":"2015"},{"key":"10.1016\/j.bspc.2026.110338_b0030","series-title":"September). Unet++: A Nested u-Net Architecture for Medical Image Segmentation","first-page":"3","author":"Zhou","year":"2018"},{"key":"10.1016\/j.bspc.2026.110338_b0035","series-title":"InProceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"4700","article-title":"Densely connected convolutional networks","author":"Huang","year":"2017"},{"key":"10.1016\/j.bspc.2026.110338_b0040","series-title":"May). Unet 3+: A Full-Scale Connected Unet for Medical Image Segmentation","first-page":"1055","author":"Huang","year":"2020"},{"key":"10.1016\/j.bspc.2026.110338_b0045","unstructured":"Oktay, O., Schlemper, J., Folgoc, L. L., Lee, M., Heinrich, M., Misawa, K., ... & Rueckert, D. (2018). Attention u-net: Learning where to look for the pancreas.arXiv preprint arXiv:1804.03999."},{"key":"10.1016\/j.bspc.2026.110338_b0050","series-title":"April). Cmu-Net: a Strong Convmixer-Based Medical Ultrasound Image Segmentation Network","first-page":"1","author":"Tang","year":"2023"},{"key":"10.1016\/j.bspc.2026.110338_b0055","unstructured":"Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., ... & Zhou, Y. (2021). Transunet: Transformers make strong encoders for medical image segmentation.arXiv preprint arXiv:2102.04306."},{"key":"10.1016\/j.bspc.2026.110338_b0060","series-title":"October). Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation","first-page":"205","author":"Cao","year":"2022"},{"issue":"5","key":"10.1016\/j.bspc.2026.110338_b0065","doi-asserted-by":"crossref","first-page":"1484","DOI":"10.1109\/TMI.2022.3230943","article-title":"Missformer: An effective transformer for 2d medical image segmentation","volume":"42","author":"Huang","year":"2022","journal-title":"IEEE Trans. Med. Imaging"},{"key":"10.1016\/j.bspc.2026.110338_b0070","series-title":"May). Mixed Transformer u-Net for Medical Image Segmentation","first-page":"2390","author":"Wang","year":"2022"},{"key":"10.1016\/j.bspc.2026.110338_b0075","series-title":"May). Cmunext: an Efficient Medical Image Segmentation Network Based on Large Kernel and Skip Fusion","first-page":"1","author":"Tang","year":"2024"},{"key":"10.1016\/j.bspc.2026.110338_b0080","series-title":"September). Utnet: a Hybrid Transformer Architecture for Medical Image Segmentation","first-page":"61","author":"Gao","year":"2021"},{"key":"10.1016\/j.bspc.2026.110338_b0085","series-title":"September). Unext: Mlp-Based Rapid Medical Image Segmentation Network","first-page":"23","author":"Valanarasu","year":"2022"},{"key":"10.1016\/j.bspc.2026.110338_b0090","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.107995","article-title":"ESDMR-Net: A lightweight network with expand-squeeze and dual multiscale residual connections for medical image segmentation","volume":"133","author":"Khan","year":"2024","journal-title":"Eng. Appl. Artif. Intel."},{"issue":"10","key":"10.1016\/j.bspc.2026.110338_b0095","doi-asserted-by":"crossref","first-page":"5574","DOI":"10.1364\/BOE.529505","article-title":"MFLUnet: multi-scale fusion lightweight Unet for medical image segmentation","volume":"15","author":"Cao","year":"2024","journal-title":"Biomed. Opt. Express"},{"key":"10.1016\/j.bspc.2026.110338_b0100","unstructured":"Ma, J., Li, F., & Wang, B. (2024). U-mamba: Enhancing long-range dependency for biomedical image segmentation.arXiv preprint arXiv:2401.04722."},{"key":"10.1016\/j.bspc.2026.110338_b0105","unstructured":"Gu, A., & Dao, T. (2023). Mamba: Linear-time sequence modeling with selective state spaces.arXiv preprint arXiv:2312.00752."},{"key":"10.1016\/j.bspc.2026.110338_b0110","unstructured":"Liao, W., Zhu, Y., Wang, X., Pan, C., Wang, Y., & Ma, L. (2024). Lightm-unet: Mamba assists in lightweight unet for medical image segmentation.arXiv preprint arXiv:2403.05246."},{"key":"10.1016\/j.bspc.2026.110338_b0115","doi-asserted-by":"crossref","first-page":"69382","DOI":"10.1109\/ACCESS.2021.3075294","article-title":"Double U-Nets for image segmentation by integrating the region and boundary information","volume":"9","author":"Guo","year":"2021","journal-title":"IEEE Access"},{"key":"10.1016\/j.bspc.2026.110338_b0120","doi-asserted-by":"crossref","DOI":"10.3389\/fgene.2021.639930","article-title":"Msu-net: Multi-scale u-net for 2d medical image segmentation","volume":"12","author":"Su","year":"2021","journal-title":"Front. Genet."},{"key":"10.1016\/j.bspc.2026.110338_b0125","doi-asserted-by":"crossref","DOI":"10.3389\/fnbot.2024.1480055","article-title":"MSGU-Net: a lightweight multi-scale ghost U-Net for image segmentation","volume":"18","author":"Cheng","year":"2025","journal-title":"Front. Neurorob."},{"key":"10.1016\/j.bspc.2026.110338_b0130","doi-asserted-by":"crossref","DOI":"10.1049\/cit2.12278","article-title":"AML\u2010Net: Attention\u2010based multi\u2010scale lightweight model for brain tumour segmentation in internet of medical things","author":"Zeeshan Aslam","year":"2024","journal-title":"CAAI Trans. Intelligence Technol., in-Press,"},{"key":"10.1016\/j.bspc.2026.110338_b0135","series-title":"InProceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3431","article-title":"Fully convolutional networks for semantic segmentation","author":"Long","year":"2015"},{"key":"10.1016\/j.bspc.2026.110338_b0140","series-title":"September). Cotr: Efficiently Bridging Cnn and Transformer for 3d Medical Image Segmentation","first-page":"171","author":"Xie","year":"2021"},{"key":"10.1016\/j.bspc.2026.110338_b0145","series-title":"InProceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"6202","article-title":"Hiformer: Hierarchical multi-scale representations using transformers for medical image segmentation","author":"Heidari","year":"2023"},{"key":"10.1016\/j.bspc.2026.110338_b0150","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2024.109353","article-title":"LATUP-Net: A lightweight 3D attention U-Net with parallel convolutions for brain tumor segmentation","volume":"184","author":"Alwadee","year":"2025","journal-title":"Comput. Biol. Med."},{"issue":"3","key":"10.1016\/j.bspc.2026.110338_b0155","doi-asserted-by":"crossref","first-page":"2088","DOI":"10.1109\/JBHI.2024.3506829","article-title":"LUCF-Net: Lightweight U-shaped cascade fusion network for medical image segmentation","volume":"29","author":"She","year":"2024","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"10.1016\/j.bspc.2026.110338_b0160","series-title":"InProceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"1042","article-title":"Mk-unet: Multi-kernel lightweight cnn for medical image segmentation","author":"Rahman","year":"2025"},{"key":"10.1016\/j.bspc.2026.110338_b0165","doi-asserted-by":"crossref","DOI":"10.1016\/j.cmpb.2025.108611","article-title":"PMFSNet: Polarized multi-scale feature self-attention network for lightweight medical image segmentation","volume":"261","author":"Zhong","year":"2025","journal-title":"Comput. Methods Programs Biomed."},{"key":"10.1016\/j.bspc.2026.110338_b0170","series-title":"December). LV-Unet: a Lightweight and Vanilla Model for Medical Image Segmentation","first-page":"4240","author":"Jiang","year":"2024"},{"key":"10.1016\/j.bspc.2026.110338_b0175","doi-asserted-by":"crossref","DOI":"10.1016\/j.cmpb.2020.105395","article-title":"DENSE-INception U-net for medical image segmentation","volume":"192","author":"Zhang","year":"2020","journal-title":"Comput. Methods Programs Biomed."},{"issue":"1","key":"10.1016\/j.bspc.2026.110338_b0180","doi-asserted-by":"crossref","first-page":"121","DOI":"10.1109\/JBHI.2020.2986926","article-title":"Multi-scale self-guided attention for medical image segmentation","volume":"25","author":"Sinha","year":"2020","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"10.1016\/j.bspc.2026.110338_b0185","series-title":"InProceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"1287","article-title":"Beyond self-attention: Deformable large kernel attention for medical image segmentation","author":"Azad","year":"2024"},{"issue":"2","key":"10.1016\/j.bspc.2026.110338_b0190","doi-asserted-by":"crossref","first-page":"699","DOI":"10.1109\/TMI.2020.3035253","article-title":"CA-Net: Comprehensive attention convolutional neural networks for explainable medical image segmentation","volume":"40","author":"Gu","year":"2020","journal-title":"IEEE Trans. Med. Imaging"},{"key":"10.1016\/j.bspc.2026.110338_b0195","series-title":"January). Sa-Unet: Spatial Attention u-Net for Retinal Vessel Segmentation","first-page":"1236","author":"Guo","year":"2021"},{"key":"10.1016\/j.bspc.2026.110338_b0200","doi-asserted-by":"crossref","unstructured":"Zhao, R., Qian, B., Zhang, X., Li, Y., Wei, R., Liu, Y., & Pan, Y. (2020, November). Rethinking dice loss for medical image segmentation. In2020 IEEE International Conference on Data Mining (ICDM)(pp. 851-860). IEEE.","DOI":"10.1109\/ICDM50108.2020.00094"},{"key":"10.1016\/j.bspc.2026.110338_b0205","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2025.108398","article-title":"Towards gastric cancer pathological segmentation: A large-scale whole-slide images dataset and dual-stream mask guided attention U-net","volume":"111","author":"Zhang","year":"2026","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.bspc.2026.110338_b0210","doi-asserted-by":"crossref","DOI":"10.1016\/j.dib.2019.104863","article-title":"Dataset of breast ultrasound images","volume":"28","author":"Al-Dhabyani","year":"2020","journal-title":"Data Brief"},{"key":"10.1016\/j.bspc.2026.110338_b0215","series-title":"September). Medical Transformer: Gated Axial-Attention for Medical Image Segmentation","first-page":"36","author":"Valanarasu","year":"2021"},{"key":"10.1016\/j.bspc.2026.110338_b0220","series-title":"September). Transfuse: Fusing Transformers and Cnns for Medical Image Segmentation","first-page":"14","author":"Zhang","year":"2021"},{"key":"10.1016\/j.bspc.2026.110338_b0225","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2023.106626","article-title":"DCSAU-Net: A deeper and more compact split-attention U-Net for medical image segmentation","volume":"154","author":"Xu","year":"2023","journal-title":"Comput. Biol. Med."},{"key":"10.1016\/j.bspc.2026.110338_b0230","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2022.102642","article-title":"WORD: A large scale dataset, benchmark and clinical applicable study for abdominal organ segmentation from CT image","volume":"82","author":"Luo","year":"2022","journal-title":"Med. Image Anal."},{"key":"10.1016\/j.bspc.2026.110338_b0235","series-title":"InProceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"684","article-title":"Lizard: a large-scale dataset for colonic nuclear instance segmentation and classification","author":"Graham","year":"2021"},{"issue":"2","key":"10.1016\/j.bspc.2026.110338_b0240","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1007\/s40747-024-01762-z","article-title":"A joint learning method for low-light facial expression recognition","volume":"11","author":"Xie","year":"2025","journal-title":"Complex Intell. Syst."},{"key":"10.1016\/j.bspc.2026.110338_b0245","article-title":"Correntropy meets cross-entropy: A robust loss against noisy labels","volume":"167","author":"Zhou","year":"2024","journal-title":"Eng. Appl. Artif. Intel."},{"issue":"1","key":"10.1016\/j.bspc.2026.110338_b0250","doi-asserted-by":"crossref","first-page":"22","DOI":"10.1007\/s10489-025-07040-w","article-title":"Bipartite graph regularized robust low-rank matrix factorization for fast semi-supervised image clustering","volume":"56","author":"Zhou","year":"2026","journal-title":"Appl. Intell."},{"issue":"1","key":"10.1016\/j.bspc.2026.110338_b0255","doi-asserted-by":"crossref","first-page":"263","DOI":"10.1109\/TCDS.2025.3590031","article-title":"BERN: A Novel framework for enhanced emotion recognition through the integration of EEG and eye movement features","volume":"18","author":"Zhou","year":"2025","journal-title":"IEEE Trans. Cognit. Dev. Syst."},{"key":"10.1016\/j.bspc.2026.110338_b0260","series-title":"InProceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"4510","article-title":"Mobilenetv2: Inverted residuals and linear bottlenecks","author":"Sandler","year":"2018"}],"container-title":["Biomedical Signal Processing and Control"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S174680942600892X?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S174680942600892X?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T23:47:06Z","timestamp":1777592826000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S174680942600892X"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,8]]},"references-count":52,"alternative-id":["S174680942600892X"],"URL":"https:\/\/doi.org\/10.1016\/j.bspc.2026.110338","relation":{},"ISSN":["1746-8094"],"issn-type":[{"value":"1746-8094","type":"print"}],"subject":[],"published":{"date-parts":[[2026,8]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"MSFLUNet: A multi-scale feature fusion lightweight U-Net for efficient medical image segmentation","name":"articletitle","label":"Article Title"},{"value":"Biomedical Signal Processing and Control","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.bspc.2026.110338","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"110338"}}