{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T01:42:33Z","timestamp":1776130953937,"version":"3.50.1"},"reference-count":59,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62471075"],"award-info":[{"award-number":["62471075"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005230","name":"Natural Science Foundation of Chongqing Municipality","doi-asserted-by":"publisher","award":["CSTB2025NSCQ-LZX0115"],"award-info":[{"award-number":["CSTB2025NSCQ-LZX0115"]}],"id":[{"id":"10.13039\/501100005230","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005230","name":"Natural Science Foundation of Chongqing Municipality","doi-asserted-by":"publisher","award":["CSTB2024NSCQ-LZX0080"],"award-info":[{"award-number":["CSTB2024NSCQ-LZX0080"]}],"id":[{"id":"10.13039\/501100005230","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005230","name":"Natural Science Foundation of Chongqing Municipality","doi-asserted-by":"publisher","award":["CSTB2023NSCQ-LZX0068"],"award-info":[{"award-number":["CSTB2023NSCQ-LZX0068"]}],"id":[{"id":"10.13039\/501100005230","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Digital Signal Processing"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1016\/j.dsp.2026.106064","type":"journal-article","created":{"date-parts":[[2026,3,16]],"date-time":"2026-03-16T07:30:38Z","timestamp":1773646238000},"page":"106064","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["ET-TransUNet: Integrating EMs-RstTrans with triple attention for robust noise suppression and feature enhancement in medical image segmentation"],"prefix":"10.1016","volume":"176","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2990-6100","authenticated-orcid":false,"given":"Qingling","family":"Xia","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4161-9910","authenticated-orcid":false,"given":"Yanjun","family":"Ren","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4359-7455","authenticated-orcid":false,"given":"Hanguang","family":"Xiao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0001-6196-2145","authenticated-orcid":false,"given":"Shihong","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9956-4501","authenticated-orcid":false,"given":"Yangjian","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4708-2245","authenticated-orcid":false,"given":"Gui-Bin","family":"Bian","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.dsp.2026.106064_bib0001","series-title":"International Conference on Digital Technologies and Applications","first-page":"974","article-title":"Medical image segmentation using deep learning: a survey","author":"Oubaalla","year":"2023"},{"key":"10.1016\/j.dsp.2026.106064_bib0002","doi-asserted-by":"crossref","first-page":"316","DOI":"10.1016\/j.inffus.2022.09.031","article-title":"Medical image segmentation using deep semantic-based methods: a review of techniques, applications and emerging trends","volume":"90","author":"Qureshi","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.106064_bib0003","series-title":"2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"3431","article-title":"Fully convolutional networks for semantic segmentation","author":"Long","year":"2015"},{"key":"10.1016\/j.dsp.2026.106064_bib0004","unstructured":"F. Yu, Multi-scale context aggregation by dilated convolutions, arXiv: 1511.07122(2015)."},{"key":"10.1016\/j.dsp.2026.106064_bib0005","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"2881","article-title":"Pyramid scene parsing network","author":"Zhao","year":"2017"},{"issue":"4","key":"10.1016\/j.dsp.2026.106064_bib0006","doi-asserted-by":"crossref","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","article-title":"DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs","volume":"40","author":"Chen","year":"2018","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2026.106064_bib0007","first-page":"3","article-title":"UNet++: a nested U-net architecture for medical image segmentation","author":"Zhou","year":"2018"},{"key":"10.1016\/j.dsp.2026.106064_bib0008","series-title":"ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"1055","article-title":"UNet 3+: a full-scale connected UNet for medical image segmentation","author":"Huang","year":"2020"},{"issue":"6","key":"10.1016\/j.dsp.2026.106064_bib0009","doi-asserted-by":"crossref","first-page":"2291","DOI":"10.1007\/s00371-022-02414-4","article-title":"SAUNet++: an automatic segmentation model of COVID-19 lesion from CT slices","volume":"39","author":"Xiao","year":"2023","journal-title":"Vis. Comput."},{"key":"10.1016\/j.dsp.2026.106064_bib0010","unstructured":"O. Oktay, J. Schlemper, L.L. Folgoc, M. Lee, M. Heinrich, K. Misawa, K. Mori, S. McDonagh, N.Y. Hammerla, B. Kainz, et al., Attention U-net: learning where to look for the pancreas, arXiv: 1804.03999(2018)."},{"key":"10.1016\/j.dsp.2026.106064_bib0011","doi-asserted-by":"crossref","first-page":"94","DOI":"10.1016\/j.isprsjprs.2020.01.013","article-title":"ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data","volume":"162","author":"Diakogiannis","year":"2020","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"10.1016\/j.dsp.2026.106064_bib0012","unstructured":"J. Chen, Y. Lu, Q. Yu, X. Luo, E. Adeli, Y. Wang, L. Lu, A.L. Yuille, Y. Zhou, TransUNet: transformers make strong encoders for medical image segmentation, arXiv: 2102.04306(2021)."},{"key":"10.1016\/j.dsp.2026.106064_bib0013","series-title":"Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"2674","article-title":"On the texture bias for few-shot CNN segmentation","author":"Azad","year":"2021"},{"key":"10.1016\/j.dsp.2026.106064_bib0014","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.104791","article-title":"Transformers in medical image segmentation: a review","volume":"84","author":"Xiao","year":"2023","journal-title":"Biomed. Signal Process. Control"},{"issue":"4","key":"10.1016\/j.dsp.2026.106064_bib0015","doi-asserted-by":"crossref","first-page":"1524","DOI":"10.1049\/cit2.12179","article-title":"Short-term and long-term memory self-attention network for segmentation of tumours in 3D medical images","volume":"8","author":"Wen","year":"2023","journal-title":"CAAI Trans. Intell. Technol."},{"key":"10.1016\/j.dsp.2026.106064_bib0016","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.dsp.2026.106064_bib0017","series-title":"Medical Image Computing and Computer Assisted Intervention\u2013MICCAI 2021: 24th International Conference, Strasbourg, France, September 27\u2013October 1, 2021, Proceedings, Part III 24","first-page":"171","article-title":"CoTr: efficiently bridging CNN and transformer for 3D medical image segmentation","author":"Xie","year":"2021"},{"key":"10.1016\/j.dsp.2026.106064_bib0018","doi-asserted-by":"crossref","DOI":"10.3389\/fbioe.2024.1398237","article-title":"DA-TransUNet: integrating spatial and channel dual attention with transformer U-net for medical image segmentation","volume":"12","author":"Sun","year":"2024","journal-title":"Front. Bioeng. Biotechnol."},{"key":"10.1016\/j.dsp.2026.106064_bib0019","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"2441","article-title":"UCTransNet: rethinking the skip connections in U-net from a channel-wise perspective with transformer","volume":"36","author":"Wang","year":"2022"},{"key":"10.1016\/j.dsp.2026.106064_bib0020","series-title":"European Conference on Computer Vision","first-page":"205","article-title":"Swin-Unet: unet-like pure transformer for medical image segmentation","author":"Cao","year":"2022"},{"key":"10.1016\/j.dsp.2026.106064_bib0021","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102634","article-title":"CSWin-UNet: transformer UNet with cross-shaped windows for medical image segmentation","volume":"113","author":"Liu","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.106064_bib0022","unstructured":"A. Dosovitskiy, An image is worth 16x16 words: transformers for image recognition at scale, arXiv: 2010.11929(2020)."},{"key":"10.1016\/j.dsp.2026.106064_bib0023","series-title":"Computer Vision \u2013 ECCV 2024","first-page":"143","article-title":"ClearCLIP: decomposing CLIP representations for dense vision-language inference","author":"Lan","year":"2025"},{"key":"10.1016\/j.dsp.2026.106064_bib0024","doi-asserted-by":"crossref","first-page":"103","DOI":"10.1016\/j.neucom.2020.12.085","article-title":"DA-DSUnet: dual attention-based dense SU-net for automatic head-and-neck tumor segmentation in MRI images","volume":"435","author":"Tang","year":"2021","journal-title":"Neurocomputing"},{"key":"10.1016\/j.dsp.2026.106064_bib0025","series-title":"2016 Fourth International Conference on 3D Vision (3DV)","first-page":"565","article-title":"V-Net: fully convolutional neural networks for volumetric medical image segmentation","author":"MilletariFausto and Navab, Nassir and Ahmadi, Seyed-Ahmad","year":"2016"},{"key":"10.1016\/j.dsp.2026.106064_bib0026","doi-asserted-by":"crossref","first-page":"74","DOI":"10.1016\/j.neunet.2019.08.025","article-title":"MultiResUNet: rethinking the U-Net architecture for multimodal biomedical image segmentation","volume":"121","author":"Ibtehaz","year":"2020","journal-title":"Neural Netw."},{"issue":"5","key":"10.1016\/j.dsp.2026.106064_bib0027","doi-asserted-by":"crossref","first-page":"389","DOI":"10.1007\/s42979-021-00782-7","article-title":"DenseNet convolutional neural networks application for predicting COVID-19 using CT image","volume":"2","author":"Hasan","year":"2021","journal-title":"SN Comput. Sci."},{"key":"10.1016\/j.dsp.2026.106064_bib0028","series-title":"ResNet 50","first-page":"63","author":"Koonce","year":"2021"},{"key":"10.1016\/j.dsp.2026.106064_bib0029","series-title":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","first-page":"128","article-title":"VA-TransUNet: a U-shaped medical image segmentation network with visual attention","author":"Jiang","year":"2022"},{"key":"10.1016\/j.dsp.2026.106064_bib0030","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2024.103280","article-title":"TransUNet: rethinking the U-Net architecture design for medical image segmentation through the lens of transformers","volume":"97","author":"Chen","year":"2024","journal-title":"Med. Image Anal."},{"key":"10.1016\/j.dsp.2026.106064_bib0031","first-page":"1","article-title":"DS-TransUNet: dual swin transformer U-net for medical image segmentation","volume":"71","author":"Lin","year":"2022","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2026.106064_bib0032","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.106472","article-title":"Enhanced-TransUNet for ultrasound segmentation of thyroid nodules","volume":"95","author":"Ozcan","year":"2024","journal-title":"Biomed. Signal Process. Control"},{"issue":"1","key":"10.1016\/j.dsp.2026.106064_bib0033","doi-asserted-by":"crossref","first-page":"85","DOI":"10.1186\/s12859-023-05196-1","article-title":"EG-TransUNet: a transformer-based U-net with enhanced and guided models for biomedical image segmentation","volume":"24","author":"Pan","year":"2023","journal-title":"BMC Bioinf."},{"key":"10.1016\/j.dsp.2026.106064_bib0034","article-title":"UCFilTransNet: cross-filtering transformer-based network for CT image segmentation","volume":"238","author":"Li","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.dsp.2026.106064_bib0035","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"10012","article-title":"Swin transformer: hierarchical vision transformer using shifted windows","author":"Liu","year":"2021"},{"issue":"1","key":"10.1016\/j.dsp.2026.106064_bib0036","doi-asserted-by":"crossref","first-page":"55","DOI":"10.1109\/TETCI.2023.3309626","article-title":"TransAttUNet: multi-level attention-guided U-net with transformer for medical image segmentation","volume":"8","author":"Chen","year":"2023","journal-title":"IEEE Trans. Emerg. Top. Comput. Intell."},{"key":"10.1016\/j.dsp.2026.106064_bib0037","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"7262","article-title":"Segmenter: transformer for semantic segmentation","author":"Strudel","year":"2021"},{"key":"10.1016\/j.dsp.2026.106064_bib0038","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102888","article-title":"A comprehensive survey of large language models and multimodal large language models in medicine","volume":"117","author":"Xiao","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.106064_bib0039","unstructured":"X. Huang, Z. Deng, D. Li, X. Yuan, MISSFormer: an effective medical image segmentation transformer, arXiv: 2109.07162(2021)."},{"key":"10.1016\/j.dsp.2026.106064_bib0040","article-title":"Multiple adverse weather image restoration: a review","volume":"618","author":"Xiao","year":"2024","journal-title":"Neurocomputing"},{"key":"10.1016\/j.dsp.2026.106064_bib0041","series-title":"Chinese Conference on Pattern Recognition and Computer Vision (PRCV)","first-page":"42","article-title":"LeViT-UNet: make faster encoders with transformer for medical image segmentation","author":"Xu","year":"2023"},{"key":"10.1016\/j.dsp.2026.106064_bib0042","first-page":"1","article-title":"UMiT-net: a U-shaped mix-transformer network for extracting precise roads using remote sensing images","volume":"61","author":"Deng","year":"2023","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"10.1016\/j.dsp.2026.106064_bib0043","series-title":"ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"2390","article-title":"Mixed transformer U-net for medical image segmentation","author":"Wang","year":"2022"},{"key":"10.1016\/j.dsp.2026.106064_bib0044","first-page":"1","article-title":"UST-net: a U-shaped transformer network using shifted windows for hyperspectral unmixing","volume":"61","author":"Yang","year":"2023","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"10.1016\/j.dsp.2026.106064_bib0045","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109131","article-title":"Mix-ViT: mixing attentive vision transformer for ultra-fine-grained visual categorization","volume":"135","author":"Yu","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.dsp.2026.106064_bib0046","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"6824","article-title":"Multiscale vision transformers","author":"Fan","year":"2021"},{"key":"10.1016\/j.dsp.2026.106064_bib0047","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"357","article-title":"CrossViT: cross-attention multi-scale vision transformer for image classification","author":"Chen","year":"2021"},{"key":"10.1016\/j.dsp.2026.106064_bib0048","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.106968","article-title":"MSFT: a multi-scale feature-based transformer model for arrhythmia classification","volume":"100","author":"Zhang","year":"2025","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.dsp.2026.106064_bib0049","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.125818","article-title":"TCMT: target-oriented cross modal transformer for multimodal aspect-based sentiment analysis","volume":"264","author":"Zou","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.dsp.2026.106064_bib0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2024.128793","article-title":"Mixed-scale cross-modal fusion network for referring image segmentation","volume":"614","author":"Pan","year":"2025","journal-title":"Neurocomputing"},{"key":"10.1016\/j.dsp.2026.106064_bib0051","series-title":"European Conference on Computer Vision","first-page":"351","article-title":"Dynamic ReLU","author":"Chen","year":"2020"},{"key":"10.1016\/j.dsp.2026.106064_bib0052","series-title":"Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18","first-page":"234","article-title":"U-Net: convolutional networks for biomedical image segmentation","author":"Ronneberger","year":"2015"},{"key":"10.1016\/j.dsp.2026.106064_bib0053","doi-asserted-by":"crossref","first-page":"55815","DOI":"10.1109\/ACCESS.2025.3555489","article-title":"CAFU-Net: a context-aware feature aggregation network for lung nodule segmentation","volume":"13","author":"Hou","year":"2025","journal-title":"IEEE Access"},{"issue":"2","key":"10.1016\/j.dsp.2026.106064_bib0054","doi-asserted-by":"crossref","first-page":"106","DOI":"10.1007\/s10044-025-01487-w","article-title":"MCT-Net: a multi-branch hybrid CNN-transformer model for medical image segmentation","volume":"28","author":"Shen","year":"2025","journal-title":"Pattern Anal. Appl."},{"key":"10.1016\/j.dsp.2026.106064_bib0055","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102634","article-title":"CSWin-UNet: transformer UNet with cross-shaped windows for medical image segmentation","volume":"113","author":"Liu","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.106064_bib0056","series-title":"Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV)","first-page":"574","article-title":"UNETR: transformers for 3D medical image segmentation","author":"Hatamizadeh","year":"2022"},{"issue":"5","key":"10.1016\/j.dsp.2026.106064_bib0057","doi-asserted-by":"crossref","first-page":"1484","DOI":"10.1109\/TMI.2022.3230943","article-title":"MISSFormer: an effective transformer for 2D medical image segmentation","volume":"42","author":"Huang","year":"2023","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"9","key":"10.1016\/j.dsp.2026.106064_bib0058","doi-asserted-by":"crossref","first-page":"4623","DOI":"10.1109\/JBHI.2022.3188710","article-title":"Full-resolution network and dual-threshold iteration for retinal vessel and coronary angiograph segmentation","volume":"26","author":"Liu","year":"2022","journal-title":"IEEE J. Biomed. Health Inf."},{"key":"10.1016\/j.dsp.2026.106064_bib0059","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"12124","article-title":"Cswin transformer: a general vision transformer backbone with cross-shaped windows","author":"Dong","year":"2022"}],"container-title":["Digital Signal Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1051200426001831?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1051200426001831?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T00:45:54Z","timestamp":1776127554000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1051200426001831"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,6]]},"references-count":59,"alternative-id":["S1051200426001831"],"URL":"https:\/\/doi.org\/10.1016\/j.dsp.2026.106064","relation":{},"ISSN":["1051-2004"],"issn-type":[{"value":"1051-2004","type":"print"}],"subject":[],"published":{"date-parts":[[2026,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"ET-TransUNet: Integrating EMs-RstTrans with triple attention for robust noise suppression and feature enhancement in medical image segmentation","name":"articletitle","label":"Article Title"},{"value":"Digital Signal Processing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.dsp.2026.106064","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Inc. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"106064"}}