{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,10]],"date-time":"2026-05-10T06:00:49Z","timestamp":1778392849214,"version":"3.51.4"},"reference-count":51,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"8","license":[{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62271220"],"award-info":[{"award-number":["62271220"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202179"],"award-info":[{"award-number":["62202179"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003819","name":"Natural Science Foundation of Hubei Province of China","doi-asserted-by":"publisher","award":["2022CFB585"],"award-info":[{"award-number":["2022CFB585"]}],"id":[{"id":"10.13039\/501100003819","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China\/Research Grants Council Joint Research Scheme","doi-asserted-by":"publisher","award":["N_HKUST627\/20"],"award-info":[{"award-number":["N_HKUST627\/20"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002920","name":"Research Grants Council","doi-asserted-by":"publisher","award":["16203319"],"award-info":[{"award-number":["16203319"]}],"id":[{"id":"10.13039\/501100002920","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Med. Imaging"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1109\/tmi.2023.3247814","type":"journal-article","created":{"date-parts":[[2023,2,22]],"date-time":"2023-02-22T18:30:48Z","timestamp":1677090648000},"page":"2325-2337","source":"Crossref","is-referenced-by-count":58,"title":["The Lighter the Better: Rethinking Transformers in Medical Image Segmentation Through Adaptive Pruning"],"prefix":"10.1109","volume":"42","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8291-4823","authenticated-orcid":false,"given":"Xian","family":"Lin","sequence":"first","affiliation":[{"name":"School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5060-2558","authenticated-orcid":false,"given":"Li","family":"Yu","sequence":"additional","affiliation":[{"name":"School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3885-4912","authenticated-orcid":false,"given":"Kwang-Ting","family":"Cheng","sequence":"additional","affiliation":[{"name":"School of Engineering, The Hong Kong University of Science and Technology, Kowloon, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2039-3863","authenticated-orcid":false,"given":"Zengqiang","family":"Yan","sequence":"additional","affiliation":[{"name":"School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan, China"}]}],"member":"263","reference":[{"key":"ref13","article-title":"TransAttUnet: Multi-level attention-guided U-Net with transformer for medical image segmentation","author":"chen","year":"2021","journal-title":"arXiv 2107 05274"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_11"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87199-4_6"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87199-4_16"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_8"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_2"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_20"},{"key":"ref16","article-title":"Medical transformer: Gated axial-attention for medical image segmentation","author":"maria jose valanarasu","year":"2021","journal-title":"arXiv 2102 10662"},{"key":"ref19","article-title":"More than encoder: Introducing transformer decoder to upsample","author":"li","year":"2021","journal-title":"arXiv 2106 10637"},{"key":"ref18","article-title":"A data-scalable transformer for medical image segmentation: Architecture, model efficiency, and benchmark","author":"gao","year":"2022","journal-title":"arXiv 2203 00131"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01171"},{"key":"ref50","article-title":"DeepViT: Towards deeper vision transformer","author":"zhou","year":"2021","journal-title":"arXiv 2103 11886"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-21014-3_39"},{"key":"ref45","article-title":"Transformers in medical image analysis: A review","author":"he","year":"2022","journal-title":"arXiv 2202 12165"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01625"},{"key":"ref47","first-page":"2286","article-title":"ConViT: Improving vision transformers with soft convolutional inductive biases","author":"d\u2019ascoli","year":"2021","journal-title":"Proc ICML"},{"key":"ref42","first-page":"1","article-title":"The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions","volume":"5","author":"tschandl","year":"2018","journal-title":"Data Science Journal"},{"key":"ref41","article-title":"Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (ISIC)","author":"codella","year":"2019","journal-title":"arXiv 1902 03368"},{"key":"ref44","article-title":"SpecTr: Spectral transformer for hyperspectral pathology image segmentation","author":"yun","year":"2021","journal-title":"arXiv 2103 03604"},{"key":"ref43","article-title":"Are convolutional neural networks or transformers more like human vision?","author":"tuli","year":"2021","journal-title":"arXiv 2105 07197"},{"key":"ref49","article-title":"Vision transformers with patch diversification","author":"gong","year":"2021","journal-title":"arXiv 2104 12753"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2019.01.012"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ITME.2018.00080"},{"key":"ref9","article-title":"TransUNet: Transformers make strong encoders for medical image segmentation","author":"chen","year":"2021","journal-title":"arXiv 2102 04306"},{"key":"ref4","first-page":"213","article-title":"End-to-end object detection with transformers","author":"carion","year":"2021","journal-title":"Proc ECCV"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref6","first-page":"234","article-title":"U-Net: Convolutional networks for biomedical image segmentation","author":"ronneberger","year":"2015","journal-title":"Proc MICCAI"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1038\/s41592-020-01008-z"},{"key":"ref35","article-title":"Swin-UNet: UNet-like pure transformer for medical image segmentation","author":"cao","year":"2021","journal-title":"arXiv 2105 05537"},{"key":"ref34","article-title":"NnFormer: Interleaved transformer for volumetric segmentation","author":"zhou","year":"2021","journal-title":"arXiv 2109 03201"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9746172"},{"key":"ref36","article-title":"LeViT-UNet: Make faster encoders with transformer for medical image segmentation","author":"xu","year":"2021","journal-title":"arXiv 2107 08623"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2020.3035253"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2019.2903562"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2020.2983721"},{"key":"ref2","article-title":"An image is worth 16&#x00D7;16 words: Transformers for image recognition at scale","author":"dosovitskiy","year":"2020","journal-title":"arXiv 2010 11929"},{"key":"ref1","first-page":"5998","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"Proc NeurIPS"},{"key":"ref39","article-title":"TransClaw U-Net: Claw U-Net with transformers for medical image segmentation","author":"chang","year":"2021","journal-title":"arXiv 2107 05188"},{"key":"ref38","article-title":"Class-aware adversarial transformers for medical image segmentation","author":"you","year":"2022","journal-title":"arXiv 2201 10737"},{"key":"ref24","article-title":"Vision transformer pruning","author":"zhu","year":"2021","journal-title":"arXiv 2104 08500"},{"key":"ref23","article-title":"D-former: A U-shaped dilated transformer for 3D medical image segmentation","author":"wu","year":"2022","journal-title":"arXiv 2201 00462"},{"key":"ref26","article-title":"DynamicViT: Efficient vision transformers with dynamic token sparsification","volume":"34","author":"rao","year":"2021","journal-title":"Proc NeurIPS"},{"key":"ref25","article-title":"IA-RED2: Interpretability-aware redundancy reduction for vision transformers","volume":"34","author":"pan","year":"2021","journal-title":"Proc NeurIPS"},{"key":"ref20","article-title":"DS-TransUNet: Dual swin transformer U-Net for medical image segmentation","author":"lin","year":"2021","journal-title":"arXiv 2106 06716"},{"key":"ref22","article-title":"MISSFormer: An effective medical image segmentation transformer","author":"huang","year":"2021","journal-title":"arXiv 2109 07162"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_31"},{"key":"ref28","first-page":"909","article-title":"Multi-label co-regularization for semi-supervised facial action unit recognition","author":"niu","year":"2019","journal-title":"Proc NeurIPS"},{"key":"ref27","article-title":"AdaViT: Adaptive tokens for efficient vision transformer","author":"yin","year":"2021","journal-title":"arXiv 2112 07658"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102327"}],"container-title":["IEEE Transactions on Medical Imaging"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/42\/10198906\/10050127.pdf?arnumber=10050127","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,8,21]],"date-time":"2023-08-21T17:52:11Z","timestamp":1692640331000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10050127\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8]]},"references-count":51,"journal-issue":{"issue":"8"},"URL":"https:\/\/doi.org\/10.1109\/tmi.2023.3247814","relation":{},"ISSN":["0278-0062","1558-254X"],"issn-type":[{"value":"0278-0062","type":"print"},{"value":"1558-254X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,8]]}}}