{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,30]],"date-time":"2025-07-30T16:06:30Z","timestamp":1753891590730,"version":"3.41.2"},"reference-count":38,"publisher":"Frontiers Media SA","license":[{"start":{"date-parts":[[2024,7,29]],"date-time":"2024-07-29T00:00:00Z","timestamp":1722211200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100020950","name":"National Science and Technology Council","doi-asserted-by":"publisher","award":["111-2311-B-007 -011 -MY3","112-2321-B-002-025 -","112-2112-M-029-003","111-2221-E-492-022-"],"award-info":[{"award-number":["111-2311-B-007 -011 -MY3","112-2321-B-002-025 -","112-2112-M-029-003","111-2221-E-492-022-"]}],"id":[{"id":"10.13039\/501100020950","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["frontiersin.org"],"crossmark-restriction":true},"short-container-title":["Front. Neuroinform."],"abstract":"<jats:p>The brain atlas, which provides information about the distribution of genes, proteins, neurons, or anatomical regions, plays a crucial role in contemporary neuroscience research. To analyze the spatial distribution of those substances based on images from different brain samples, we often need to warp and register individual brain images to a standard brain template. However, the process of warping and registration may lead to spatial errors, thereby severely reducing the accuracy of the analysis. To address this issue, we develop an automated method for segmenting neuropils in the <jats:italic>Drosophila<\/jats:italic> brain for fluorescence images from the <jats:italic>FlyCircuit<\/jats:italic> database. This technique allows future brain atlas studies to be conducted accurately at the individual level without warping and aligning to a standard brain template. Our method, LYNSU (Locating by YOLO and Segmenting by U-Net), consists of two stages. In the first stage, we use the YOLOv7 model to quickly locate neuropils and rapidly extract small-scale 3D images as input for the second stage model. This stage achieves a 99.4% accuracy rate in neuropil localization. In the second stage, we employ the 3D U-Net model to segment neuropils. LYNSU can achieve high accuracy in segmentation using a small training set consisting of images from merely 16 brains. We demonstrate LYNSU on six distinct neuropils or structures, achieving a high segmentation accuracy comparable to professional manual annotations with a 3D Intersection-over-Union (IoU) reaching up to 0.869. Our method takes only about 7 s to segment a neuropil while achieving a similar level of performance as the human annotators. To demonstrate a use case of LYNSU, we applied it to all female <jats:italic>Drosophila<\/jats:italic> brains from the <jats:italic>FlyCircuit<\/jats:italic> database to investigate the asymmetry of the mushroom bodies (MBs), the learning center of fruit flies. We used LYNSU to segment bilateral MBs and compare the volumes between left and right for each individual. Notably, of 8,703 valid brain samples, 10.14% showed bilateral volume differences that exceeded 10%. The study demonstrated the potential of the proposed method in high-throughput anatomical analysis and connectomics construction of the <jats:italic>Drosophila<\/jats:italic> brain.<\/jats:p>","DOI":"10.3389\/fninf.2024.1429670","type":"journal-article","created":{"date-parts":[[2024,7,29]],"date-time":"2024-07-29T04:31:02Z","timestamp":1722227462000},"update-policy":"https:\/\/doi.org\/10.3389\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["LYNSU: automated 3D neuropil segmentation of fluorescent images for Drosophila brains"],"prefix":"10.3389","volume":"18","author":[{"given":"Kai-Yi","family":"Hsu","sequence":"first","affiliation":[]},{"given":"Chi-Tin","family":"Shih","sequence":"additional","affiliation":[]},{"given":"Nan-Yow","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Chung-Chuan","family":"Lo","sequence":"additional","affiliation":[]}],"member":"1965","published-online":{"date-parts":[[2024,7,29]]},"reference":[{"key":"B1","doi-asserted-by":"publisher","DOI":"10.1038\/s42003-023-05468-9","article-title":"Imagining the future of optical microscopy: Everything, everywhere, all at once.","volume":"6","author":"Balasubramanian","year":"2023","journal-title":"Commun. Biol."},{"key":"B2","doi-asserted-by":"publisher","first-page":"524","DOI":"10.1038\/s41583-020-0363-6","article-title":"Reflections on the past two decades of neuroscience.","volume":"21","author":"Bassett","year":"2020","journal-title":"Nat. Rev. Neurosci."},{"key":"B3","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2003.01995","article-title":"A learning strategy for contrast-agnostic MRI segmentation.","author":"Billot","year":"2020","journal-title":"arXiv"},{"key":"B4","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0236495","article-title":"An unbiased template of the Drosophila brain and ventral nerve cord.","volume":"15","author":"Bogovic","year":"2020","journal-title":"PLoS One"},{"key":"B5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.cub.2010.11.056","article-title":"Three-Dimensional reconstruction of brain-wide wiring networks in Drosophila at single-cell resolution.","volume":"21","author":"Chiang","year":"2011","journal-title":"Curr. Biol."},{"key":"B6","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-019-12715-3","article-title":"Rapid single-wavelength lightsheet localization microscopy for clarified tissue.","volume":"10","author":"Chu","year":"2019","journal-title":"Nat. Commun."},{"key":"B7","doi-asserted-by":"publisher","DOI":"10.3389\/fphys.2023.1076533","article-title":"Virtual Fly Brain\u2014An interactive atlas of the Drosophila nervous system.","volume":"14","author":"Court","year":"2023","journal-title":"Front. Physiol."},{"key":"B8","doi-asserted-by":"publisher","first-page":"456","DOI":"10.1016\/j.neuroimage.2017.04.039","article-title":"3D fully convolutional networks for subcortical segmentation in MRI: A large-scale study.","volume":"170","author":"Dolz","year":"2018","journal-title":"Neuroimage"},{"key":"B9","doi-asserted-by":"publisher","first-page":"584","DOI":"10.1039\/D1LC00827G","article-title":"Development of a water refractive index-matched microneedle integrated into a light sheet microscopy system for continuous embryonic cell imaging.","volume":"22","author":"Fan","year":"2022","journal-title":"Lab. Chip"},{"key":"B10","doi-asserted-by":"publisher","DOI":"10.1364\/OL.428562","article-title":"Light field microscopy based on structured light illumination.","volume":"46","author":"Fu","year":"2021","journal-title":"Opt. Lett."},{"key":"B11","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1512.03385","article-title":"Deep residual learning for image recognition.","author":"He","year":"2015","journal-title":"arXiv"},{"key":"B12","doi-asserted-by":"publisher","DOI":"10.1186\/s12915-017-0461-8","article-title":"Q&A: Why use synchrotron x-ray tomography for multi-scale connectome mapping?","volume":"15","author":"Hwu","year":"2017","journal-title":"BMC Biol."},{"journal-title":"Segmentation models. GitHub.","year":"2019","author":"Iakubovskii","key":"B13"},{"key":"B14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2024\/3819801","article-title":"YOLO-UNet architecture for detecting and segmenting the localized MRI brain tumor image.","volume":"2024","author":"Iriawan","year":"2024","journal-title":"Appl. Comput. Intell. Soft Comput."},{"key":"B15","doi-asserted-by":"publisher","first-page":"1187","DOI":"10.1016\/j.cell.2007.01.040","article-title":"Comprehensive maps of drosophila higher olfactory centers: Spatially segregated fruit and pheromone representation.","volume":"128","author":"Jefferis","year":"2007","journal-title":"Cell"},{"key":"B16","doi-asserted-by":"publisher","first-page":"284","DOI":"10.1016\/j.celrep.2019.12.018","article-title":"Neurochemical Organization of the drosophila brain visualized by endogenously tagged neurotransmitter receptors.","volume":"30","author":"Kondo","year":"2020","journal-title":"Cell Rep."},{"key":"B17","doi-asserted-by":"publisher","first-page":"9042","DOI":"10.1523\/JNEUROSCI.1527-18.2019","article-title":"Wide. Fast. Deep: Recent advances in multiphoton microscopy of in vivo neuronal activity.","volume":"39","author":"Lecoq","year":"2019","journal-title":"J. Neurosci."},{"key":"B18","doi-asserted-by":"publisher","DOI":"10.1038\/srep08920","article-title":"Visualization of synaptic domains in the Drosophila brain by magnetic resonance microscopy at 10 micron isotropic resolution.","volume":"5","author":"Lee","year":"2015","journal-title":"Sci. Rep."},{"key":"B19","doi-asserted-by":"publisher","first-page":"1205","DOI":"10.1016\/j.cell.2007.03.006","article-title":"A map of olfactory representation in the drosophila mushroom body.","volume":"128","author":"Lin","year":"2007","journal-title":"Cell"},{"key":"B20","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1016\/j.isci.2019.03.025","article-title":"Imaging through the Whole brain of Drosophila at \u03bb\/20 super-resolution.","volume":"14","author":"Lin","year":"2019","journal-title":"iScience"},{"key":"B21","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1411.4038","article-title":"Fully convolutional networks for semantic segmentation.","author":"Long","year":"2014","journal-title":"arXiv"},{"key":"B22","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1162\/netn_a_00283","article-title":"Circuit analysis of the Drosophila brain using connectivity-based neuronal classification reveals organization of key communication pathways.","volume":"7","author":"Mehta","year":"2023","journal-title":"Netw. Neurosci."},{"key":"B23","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-023-33289-7","article-title":"Automatic brain extraction and brain tissues segmentation on multi-contrast animal MRI.","volume":"13","author":"Nour Eddin","year":"2023","journal-title":"Sci. Rep."},{"key":"B24","doi-asserted-by":"publisher","DOI":"10.1038\/s43586-022-00131-9","article-title":"Volume electron microscopy.","volume":"2","author":"Peddie","year":"2022","journal-title":"Nat. Rev. Methods Prim."},{"key":"B25","doi-asserted-by":"publisher","first-page":"493","DOI":"10.1038\/nmeth.1602","article-title":"BrainAligner: 3D registration atlases of Drosophila brains.","volume":"8","author":"Peng","year":"2011","journal-title":"Nat. Methods"},{"key":"B26","doi-asserted-by":"publisher","first-page":"367","DOI":"10.1038\/s41587-022-01471-3","article-title":"Rationalized deep learning super-resolution microscopy for sustained live imaging of rapid subcellular processes.","volume":"41","author":"Qiao","year":"2023","journal-title":"Nat. Biotechnol."},{"key":"B27","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-021-90428-8","article-title":"Brain tumor segmentation based on deep learning and an attention mechanism using MRI multi-modalities brain images.","volume":"11","author":"Ranjbarzadeh","year":"2021","journal-title":"Sci. Rep."},{"key":"B28","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1505.04597","article-title":"U-Net: Convolutional networks for biomedical image segmentation.","author":"Ronneberger","year":"2015","journal-title":"arXiv"},{"key":"B29","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1801.04161","article-title":"QuickNAT: A fully convolutional network for quick and accurate segmentation of neuroanatomy.","author":"Roy","year":"2018","journal-title":"arXiv"},{"key":"B30","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2021.105089","article-title":"3D convolutional neural networks for stalled brain capillary detection.","volume":"141","author":"Solovyev","year":"2022","journal-title":"Comput. Biol. Med."},{"key":"B31","doi-asserted-by":"publisher","DOI":"10.3389\/fnins.2020.00179","article-title":"DeepBrainSeg: Automated brain region segmentation for micro-optical images with a convolutional neural network.","volume":"14","author":"Tan","year":"2020","journal-title":"Front. Neurosci."},{"key":"B32","doi-asserted-by":"publisher","first-page":"1522","DOI":"10.1038\/s41593-020-00739-8","article-title":"Recent advances in neurotechnologies with broad potential for neuroscience research.","volume":"23","author":"V\u00e1zquez-Guardado","year":"2020","journal-title":"Nat. Neurosci."},{"key":"B33","doi-asserted-by":"publisher","DOI":"10.1186\/s12880-022-00807-4","article-title":"Deep learning-based automated segmentation of eight brain anatomical regions using head CT images in PET\/CT.","volume":"22","author":"Wang","year":"2022","journal-title":"BMC Med. Imaging"},{"key":"B34","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2207.02696","article-title":"YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors.","author":"Wang","year":"2022","journal-title":"arXiv"},{"key":"B35","doi-asserted-by":"publisher","DOI":"10.1186\/s40478-023-01509-w","article-title":"AI-based MRI auto-segmentation of brain tumor in rodents, a multicenter study.","volume":"11","author":"Wang","year":"2023","journal-title":"Acta Neuropathol. Commun."},{"key":"B36","doi-asserted-by":"publisher","first-page":"11971","DOI":"10.1021\/acs.chemrev.0c01174","article-title":"Super-resolving microscopy in neuroscience.","volume":"121","author":"Werner","year":"2021","journal-title":"Chem. Rev."},{"key":"B37","doi-asserted-by":"publisher","first-page":"2585","DOI":"10.1002\/cne.24512","article-title":"Neuroarchitecture of the Drosophila central complex: A catalog of nodulus and asymmetrical body neurons and a revision of the protocerebral bridge catalog.","volume":"526","author":"Wolff","year":"2018","journal-title":"J. Comp. Neurol."},{"key":"B38","doi-asserted-by":"publisher","DOI":"10.3389\/fneur.2021.653375","article-title":"Multi-modal segmentation of 3D brain scans using neural networks.","volume":"12","author":"Zopes","year":"2021","journal-title":"Front. Neurol."}],"container-title":["Frontiers in Neuroinformatics"],"original-title":[],"link":[{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/fninf.2024.1429670\/full","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,29]],"date-time":"2024-07-29T04:31:11Z","timestamp":1722227471000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/fninf.2024.1429670\/full"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,29]]},"references-count":38,"alternative-id":["10.3389\/fninf.2024.1429670"],"URL":"https:\/\/doi.org\/10.3389\/fninf.2024.1429670","relation":{},"ISSN":["1662-5196"],"issn-type":[{"type":"electronic","value":"1662-5196"}],"subject":[],"published":{"date-parts":[[2024,7,29]]},"article-number":"1429670"}}