{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,29]],"date-time":"2025-03-29T04:16:52Z","timestamp":1743221812245,"version":"3.40.3"},"reference-count":48,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100000646","name":"Japan Society for the Promotion of Science (JSPS) KAKENHI","doi-asserted-by":"publisher","award":["24K15020"],"award-info":[{"award-number":["24K15020"]}],"id":[{"id":"10.13039\/501100000646","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3552847","type":"journal-article","created":{"date-parts":[[2025,3,19]],"date-time":"2025-03-19T19:48:13Z","timestamp":1742413693000},"page":"50918-50930","source":"Crossref","is-referenced-by-count":0,"title":["Human Brain-Inspired Network Using Transformer and Feedback Processing for Cell Image Segmentation"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-6969-4017","authenticated-orcid":false,"given":"Hinako","family":"Mitsuoka","sequence":"first","affiliation":[{"name":"Department of Electrical and Electronic Engineering, Meijo University, Nagoya, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5675-8713","authenticated-orcid":false,"given":"Kazuhiro","family":"Hotta","sequence":"additional","affiliation":[{"name":"Department of Electrical and Electronic Engineering, Meijo University, Nagoya, Japan"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/3DV.2016.79"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ISBI.2019.8759447"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1038\/s41592-021-01249-6"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2644615"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.660"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.2983686"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2017.2699184"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref10","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv:2010.11929"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00621"},{"key":"ref13","article-title":"ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness","author":"Geirhos","year":"2018","journal-title":"arXiv:1811.12231"},{"key":"ref14","article-title":"Are convolutional neural networks or transformers more like human vision?","author":"Tuli","year":"2021","journal-title":"arXiv:2105.07197"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00420"},{"issue":"11","key":"ref16","doi-asserted-by":"crossref","first-page":"571","DOI":"10.1016\/S0166-2236(00)01657-X","article-title":"The distinct modes of vision offered by feedforward and recurrent processing","volume":"23","author":"Lamme","year":"2000","journal-title":"Trends Neurosciences"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/S0960-9822(03)00337-3"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1093\/cercor\/1.1.1"},{"key":"ref19","first-page":"719","volume-title":"Perceptrons, Adalines, and Backpropagation","author":"Widrow","year":"1998"},{"key":"ref20","first-page":"1","article-title":"Accuracy improvement of cell image segmentation using feedback former","volume-title":"Proc. Comput. Vis. Workshops","author":"Mitsuoka"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01055"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2023.3329173"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00656"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-66415-2_24"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2019.00145"},{"key":"ref27","article-title":"Rethinking Atrous convolution for semantic image segmentation","author":"Chen","year":"2017","journal-title":"arXiv:1706.05587"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1802.02611"},{"key":"ref29","first-page":"12077","article-title":"SegFormer: Simple and efficient design for semantic segmentation with transformers","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Xie"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00135"},{"key":"ref32","article-title":"Recurrent residual convolutional neural network based on U-Net (R2U-Net) for medical image segmentation","author":"Alom","year":"2018","journal-title":"arXiv:1802.06955"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.3934\/mbe.2023535"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2021.3060497"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00172"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298958"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-50835-1_22"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01871"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref40","first-page":"15909","article-title":"RepViT: Revisiting mobile CNN from ViT perspective","volume-title":"Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit.","author":"Wang"},{"key":"ref41","article-title":"Segmented anisotropic ssTEM dataset of neural tissue","author":"Gerhard","year":"2013","journal-title":"Figshare"},{"volume-title":"Segmentation of Neuronal Structures in EM Stacks Challenge","year":"2012","key":"ref42"},{"key":"ref43","article-title":"SGDR: Stochastic gradient descent with warm restarts","author":"Loshchilov","year":"2016","journal-title":"arXiv:1608.03983"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00475"},{"key":"ref45","first-page":"3965","article-title":"CoAtNet: Marrying convolution and attention for all data sizes","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Dai"},{"issue":"11","key":"ref46","doi-asserted-by":"crossref","first-page":"3791","DOI":"10.1007\/s00371-021-02221-3","article-title":"Cell image segmentation by using feedback and convolutional LSTM","volume":"38","author":"Shibuya","year":"2022","journal-title":"Vis. Comput."},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2102.04306"},{"key":"ref48","article-title":"BEFUNet: A hybrid CNN-transformer architecture for precise medical image segmentation","author":"Manzari","year":"2024","journal-title":"arXiv:2402.08793"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/10934004.pdf?arnumber=10934004","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,29]],"date-time":"2025-03-29T03:53:49Z","timestamp":1743220429000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10934004\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":48,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3552847","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2025]]}}}