{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T05:47:59Z","timestamp":1767332879845,"version":"3.48.0"},"reference-count":46,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2025203025"],"award-info":[{"award-number":["F2025203025"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100015973","name":"Basic Innovative Research Cultivation Program of Yanshan University","doi-asserted-by":"publisher","award":["2024LGQN009"],"award-info":[{"award-number":["2024LGQN009"]}],"id":[{"id":"10.13039\/501100015973","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3647946","type":"journal-article","created":{"date-parts":[[2025,12,24]],"date-time":"2025-12-24T18:46:53Z","timestamp":1766602013000},"page":"217349-217363","source":"Crossref","is-referenced-by-count":0,"title":["EECD-Net: An Energy-Efficient Crack Detection Framework Integrating Super-Resolution, Spiking Neural Networks, and Gated Attention"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-0819-4409","authenticated-orcid":false,"given":"Shuo","family":"Zhang","sequence":"first","affiliation":[{"name":"School of Information Science and Engineering, Yanshan University, Qinhuangdao, China"}]},{"given":"Shuai","family":"Tong","sequence":"additional","affiliation":[{"name":"College of Science, Harbin University of Science and Technology, Harbin, Heilongjiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2354-2577","authenticated-orcid":false,"given":"Kuo","family":"Pang","sequence":"additional","affiliation":[{"name":"School of Information Science and Engineering, Yanshan University, Qinhuangdao, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.3390\/rs16162910"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/s41062-025-02147-y"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-025-20729-x"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2878966"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3003638"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2025.3540841"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2023.105112"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2024.3403389"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2024.105960"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-025-01065-1"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2025.110178"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i1.27816"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2025.110227"},{"key":"ref14","article-title":"MobileNets: Efficient convolutional neural networks for mobile vision applications","author":"Howard","year":"2017","journal-title":"arXiv:1704.04861"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00716"},{"key":"ref16","first-page":"10096","article-title":"EfficientNetV2: Smaller models and faster training","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Tan"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00140"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00286"},{"key":"ref19","first-page":"12077","article-title":"SegFormer: Simple and efficient design for semantic segmentation with transformers","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Xie"},{"key":"ref20","article-title":"MobileViT: Light-weight, general-purpose, and mobile-friendly vision transformer","author":"Mehta","year":"2021","journal-title":"arXiv:2110.02178"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ISSCC.2014.6757323"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2021\/321"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2024.111681"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2022.3179968"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2025.129821"},{"key":"ref26","first-page":"11173","article-title":"Event-based object detection via posterior fusion","volume-title":"Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit.","author":"Li"},{"key":"ref27","article-title":"Oriented feature alignment for fine-grained object recognition in high-resolution satellite imagery","author":"Ming","year":"2021","journal-title":"arXiv:2110.06628"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1809.00219"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.23919\/MVA51890.2021.9511400"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.3390\/s22239092"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01357"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2023.105047"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2025.113027"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1088\/1361-6501\/addc04"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.3390\/brainsci12070863"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2025.129664"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2023\/342"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TII.2024.3366946"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1016\/j.istruc.2025.108720"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-73411-3_15"},{"key":"ref42","article-title":"RobustVLA: Robustness-aware reinforcement post-training for vision-language-action models","author":"Zhang","year":"2025","journal-title":"arXiv:2511.01331"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref44","article-title":"Attention U-Net: Learning where to look for the pancreas","author":"Oktay","year":"2018","journal-title":"arXiv:1804.03999"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2102.04306"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/11314167.pdf?arnumber=11314167","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T05:45:40Z","timestamp":1767332740000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11314167\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":46,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3647946","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2025]]}}}