{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,20]],"date-time":"2026-04-20T18:40:54Z","timestamp":1776710454480,"version":"3.51.2"},"reference-count":66,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/100005825","name":"United States Department of Agriculture (USDA) National Institute of Food and Agriculture","doi-asserted-by":"publisher","award":["2022-67021-38911"],"award-info":[{"award-number":["2022-67021-38911"]}],"id":[{"id":"10.13039\/100005825","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/access.2024.3451304","type":"journal-article","created":{"date-parts":[[2024,8,28]],"date-time":"2024-08-28T19:00:23Z","timestamp":1724871623000},"page":"122139-122154","source":"Crossref","is-referenced-by-count":30,"title":["seUNet-Trans: A Simple Yet Effective UNet-Transformer Model for Medical Image Segmentation"],"prefix":"10.1109","volume":"12","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-0499-3882","authenticated-orcid":false,"given":"Tan-Hanh","family":"Pham","sequence":"first","affiliation":[{"name":"Department of Mechanical Engineering, Florida Institute of Technology, Melbourne, FL, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7687-4287","authenticated-orcid":false,"given":"Xianqi","family":"Li","sequence":"additional","affiliation":[{"name":"Department of Mathematics and Systems Engineering, Florida Institute of Technology, Melbourne, FL, USA"}]},{"given":"Kim-Doang","family":"Nguyen","sequence":"additional","affiliation":[{"name":"Department of Mechanical Engineering, Florida Institute of Technology, Melbourne, FL, USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-018-0107-6"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-020-2145-8"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/s10278-019-00227-x"},{"key":"ref6","article-title":"Automated design of deep learning methods for biomedical image segmentation","author":"Isensee","year":"2019","journal-title":"arXiv:1904.08128"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1038\/s41592-020-01008-z"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2019.01.012"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1407"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.476"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_4"},{"key":"ref13","article-title":"Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer","author":"Zagoruyko","year":"2016","journal-title":"arXiv:1612.03928"},{"key":"ref14","article-title":"Rethinking atrous convolution for semantic image segmentation","author":"Chen","year":"2017","journal-title":"arXiv:1706.05587"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref16","article-title":"An image is worth 16 \u00d7 16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv:2010.11929"},{"key":"ref17","article-title":"ImageNet classification with deep convolutional neural networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"25","author":"Krizhevsky"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"ref20","article-title":"Deformable DETR: Deformable transformers for end-to-end object detection","author":"Zhu","year":"2020","journal-title":"arXiv:2010.04159"},{"key":"ref21","first-page":"15908","article-title":"Transformer in transformer","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Han"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01716"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2024.108650"},{"key":"ref24","first-page":"9355","article-title":"Twins: Revisiting the design of spatial attention in vision transformers","volume-title":"Proc. 35th Conf. Neural Inf. Process. Syst.","volume":"34","author":"Chu"},{"key":"ref25","article-title":"Shuffle transformer: Rethinking spatial shuffle for vision transformer","author":"Huang","year":"2021","journal-title":"arXiv:2106.03650"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3206108"},{"key":"ref27","article-title":"Visual parser: Representing part-whole hierarchies with transformers","author":"Sun","year":"2021","journal-title":"arXiv:2107.05790"},{"key":"ref28","article-title":"Toward transformer-based object detection","author":"Beal","year":"2020","journal-title":"arXiv:2012.09958"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2102.04306"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-87193-2_2"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"ref32","article-title":"Attention U-Net: Learning where to look for the pancreas","author":"Oktay","year":"2018","journal-title":"arXiv:1804.03999"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/j.isprsjprs.2020.01.013"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref35","article-title":"Recurrent residual convolutional neural network based on U-Net (R2U-Net) for medical image segmentation","author":"Zahangir Alom","year":"2018","journal-title":"arXiv:1802.06955"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1802.00368"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-59719-1_36"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CBMS49503.2020.00111"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3159394"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00717"},{"key":"ref41","first-page":"12077","article-title":"SegFormer: Simple and efficient design for semantic segmentation with transformers","volume-title":"Proc. Adv. Neural Inf. Process. Sys. (NIPS)","volume":"34","author":"Xie"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2022.3178991"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/tmi.2024.3398728"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2023.3293771"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109228"},{"key":"ref48","first-page":"255","article-title":"Convolutional networks for images, speech, and time series","volume-title":"The Handbook of Brain Theory and Neural Networks","author":"LeCu","year":"1995"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.5555\/3045118.3045167"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.5555\/3104322.3104425"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/ICESC51422.2021.9532631"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"ref53","article-title":"Gaussian error linear units (GELUs)","author":"Hendrycks","year":"2016","journal-title":"arXiv:1606.08415"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3063716"},{"key":"ref55","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014","journal-title":"arXiv:1412.6980"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/ITME.2018.00080"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/ISM46123.2019.00049"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.660"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.2983686"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1802.02611"},{"key":"ref61","article-title":"HarDNet-MSEG: A simple encoder\u2013decoder polyp segmentation neural network that achieves over 0.9 mean dice and 86 FPS","author":"Huang","year":"2021","journal-title":"arXiv:2101.07172"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-32239-7_34"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/LGRS.2018.2802944"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2023.106626"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2644615"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00052"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10380310\/10654259.pdf?arnumber=10654259","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T04:59:40Z","timestamp":1726030780000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10654259\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":66,"URL":"https:\/\/doi.org\/10.1109\/access.2024.3451304","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]}}}