{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T11:01:55Z","timestamp":1740135715776,"version":"3.37.3"},"reference-count":29,"publisher":"Cambridge University Press (CUP)","license":[{"start":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T00:00:00Z","timestamp":1734307200000},"content-version":"unspecified","delay-in-days":350,"URL":"https:\/\/www.cambridge.org\/core\/terms"}],"funder":[{"DOI":"10.13039\/501100017668","name":"Anhui Provincial Key Research and Development Plan","doi-asserted-by":"publisher","award":["202304A05020063"],"award-info":[{"award-number":["202304A05020063"]}],"id":[{"id":"10.13039\/501100017668","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["cambridge.org"],"crossmark-restriction":true},"short-container-title":["AIEDAM"],"published-print":{"date-parts":[[2024]]},"abstract":"<jats:title>Abstract<\/jats:title>\n\t  <jats:p>Bridge engineering design drawings basic elements contain a large amount of important information such as structural dimensions and material indexes. Basic element detection is seen as the basis for digitizing drawings. Aiming at the problem of low detection accuracy of existing drawing basic elements, an improved basic elements detection algorithm for bridge engineering design drawings based on YOLOv5 is proposed. Firstly, coordinate attention is introduced into the feature extraction network to enhance the feature extraction capability of the algorithm and alleviate the problem of difficult recognition of texture features inside grayscale images. Then, targeting objectives across different scales, the standard 3\u00a0\u00d7\u00a03 convolution in the feature pyramid network is replaced with switchable atrous convolution, and the atrous rate is adaptively selected for convolution computation to expand the sensory field. Finally, experiments are conducted on the bridge engineering design drawings basic elements detection dataset, and the experimental results show that when the Intersection over Union is 0.5, the proposed algorithm achieves a mean average precision of 93.6%, which is 3.4% higher compared to the original YOLOv5 algorithm, and it can satisfy the accuracy requirement of bridge engineering design drawings basic elements detection.<\/jats:p>","DOI":"10.1017\/s089006042400026x","type":"journal-article","created":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T14:15:02Z","timestamp":1734358502000},"update-policy":"https:\/\/doi.org\/10.1017\/policypage","source":"Crossref","is-referenced-by-count":0,"title":["Improved basic elements detection algorithm for bridge engineering design drawings based on YOLOv5"],"prefix":"10.1017","volume":"38","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-0884-1648","authenticated-orcid":false,"given":"Ning","family":"An","sequence":"first","affiliation":[]},{"given":"Linsheng","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Mengnan","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Junan","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Chuanjian","family":"Wang","sequence":"additional","affiliation":[]}],"member":"56","published-online":{"date-parts":[[2024,12,16]]},"reference":[{"key":"S089006042400026X_r21","first-page":"6","article-title":"Topology-based engineering drawing recognition method","volume":"27","author":"Song","year":"2011","journal-title":"Journal of Shenyang University of Architecture (Natural Science Edition)"},{"key":"S089006042400026X_r24","doi-asserted-by":"crossref","unstructured":"Woo, S , Park, J , Lee, J-Y and Kweon, I-S . 2018. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (Eccv), 3\u201319.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"S089006042400026X_r14","doi-asserted-by":"crossref","unstructured":"Lin, T-Y , Doll\u00e1r, P , Girshick, R , He, K , Hariharan, B and Belongie, S . 2017. Feature pyramid networks for object detection. In Proceedings of the Ieee Conference on Computer Vision and Pattern Recognition, 2117\u20132125.","DOI":"10.1109\/CVPR.2017.106"},{"key":"S089006042400026X_r16","doi-asserted-by":"crossref","unstructured":"Liu, S , Qi, L , Qin, H , Shi, J and Jia, J . 2018. Path aggregation network for instance segmentation. In Proceedings of the Ieee Conference on Computer Vision and Pattern Recognition, 8759\u20138768.","DOI":"10.1109\/CVPR.2018.00913"},{"key":"S089006042400026X_r11","first-page":"1","volume-title":"2020 International Joint Conference on Neural Networks (Ijcnn)","author":"Jamieson","year":"2020"},{"volume-title":"37th Computers and Information in Engineering Conference","year":"2017","author":"Brock","key":"S089006042400026X_r1"},{"key":"S089006042400026X_r5","doi-asserted-by":"publisher","DOI":"10.1038\/nature21056"},{"key":"S089006042400026X_r10","unstructured":"Jaderberg, M , Simonyan, K , Zisserman, A , et al. 2015. Spatial transformer networks. Advances in Neural Information Processing Systems 28."},{"key":"S089006042400026X_r30","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6999"},{"key":"S089006042400026X_r13","doi-asserted-by":"publisher","DOI":"10.1109\/MCOM.001.21664"},{"key":"S089006042400026X_r29","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2021.103750"},{"key":"S089006042400026X_r15","doi-asserted-by":"crossref","unstructured":"Liu, L , Chen, Y and Liu, X . 2019. Engineering drawing recognition model with convolutional neural network. Proceedings of the 2019 International Conference on Robotics, Intelligent Control and Artificial Intelligence (RICAI 2019), 112\u2013116.","DOI":"10.1145\/3366194.3366213"},{"key":"S089006042400026X_r28","first-page":"4793","article-title":"Deep hough transform for semantic line detection","volume":"44","author":"Zhao","year":"2022","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"S089006042400026X_r22","doi-asserted-by":"crossref","unstructured":"Wang, C-Y , Mark Liao, H-Y , Wu, Y-H , Chen, P-Y , Hsieh, J-W and Yeh, I-H . 2020. Cspnet: A new backbone that can enhance learning capability of cnn. In Proceedings of the Ieee\/Cvf Conference on Computer Vision and Pattern Recognition Workshops, 390\u2013391.","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"S089006042400026X_r20","doi-asserted-by":"crossref","unstructured":"Son, H , Lee, J , Cho, S and Lee, S . 2021. Single image defocus deblurring using kernel-sharing parallel atrous convolutions. In Proceedings of the Ieee\/Cvf International Conference on Computer Vision, 2642\u20132650.","DOI":"10.1109\/ICCV48922.2021.00264"},{"key":"S089006042400026X_r7","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2389824"},{"key":"S089006042400026X_r27","doi-asserted-by":"publisher","DOI":"10.1145\/3158369"},{"key":"S089006042400026X_r25","first-page":"39","article-title":"Two-dimensional drawing recognition of duct planes based on improved cascade rcnn","volume":"4","author":"Yang","year":"2022","journal-title":"Journal of Civil Engineering and Management"},{"key":"S089006042400026X_r12","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2023.3275708"},{"key":"S089006042400026X_r2","doi-asserted-by":"publisher","DOI":"10.1109\/TCE.2023.3257201"},{"key":"S089006042400026X_r8","doi-asserted-by":"crossref","unstructured":"Hou, Q , Zhou, D and Feng, J . 2021. Coordinate attention for efficient mobile network design. In Proceedings of the Ieee\/Cvf Conference on Computer Vision and Pattern Recognition, 13713\u201313722.","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"S089006042400026X_r17","doi-asserted-by":"publisher","DOI":"10.3390\/e22080840"},{"key":"S089006042400026X_r3","first-page":"27","article-title":"Character recognition and detection algorithm for power grid engineering drawings based on improved convolutional neural network","volume":"31","author":"Dong","year":"2023b","journal-title":"Electronic Design Engineering"},{"key":"S089006042400026X_r23","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-022-08077-5"},{"key":"S089006042400026X_r4","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2020.05.025"},{"key":"S089006042400026X_r9","doi-asserted-by":"crossref","unstructured":"Hu, J , Shen, L and Sun, G . 2018. Squeeze-and-excitation networks. In Proceedings of the Ieee Conference on Computer Vision and Pattern Recognition, 7132\u20137141.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"S089006042400026X_r18","doi-asserted-by":"crossref","unstructured":"Qiao, S , Chen, L-C and Yuille, A . 2021. Detectors: Detecting objects with recursive feature pyramid and switchable atrous convolution. In Proceedings of the Ieee\/Cvf Conference on Computer Vision and Pattern Recognition, 10213\u201310224.","DOI":"10.1109\/CVPR46437.2021.01008"},{"key":"S089006042400026X_r19","doi-asserted-by":"crossref","unstructured":"Selvaraju, RR , Cogswell, M , Das, A , Vedantam, R , Parikh, D and Batra, D . 2017. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the Ieee International Conference on Computer Vision, 618\u2013626.","DOI":"10.1109\/ICCV.2017.74"},{"key":"S089006042400026X_r6","first-page":"161","article-title":"Engineering drawing string and labeling information extraction","volume":"48","author":"Fan","year":"2012","journal-title":"Computer Engineering and Application"}],"container-title":["Artificial Intelligence for Engineering Design, Analysis and Manufacturing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.cambridge.org\/core\/services\/aop-cambridge-core\/content\/view\/S089006042400026X","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T14:15:15Z","timestamp":1734358515000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.cambridge.org\/core\/product\/identifier\/S089006042400026X\/type\/journal_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":29,"alternative-id":["S089006042400026X"],"URL":"https:\/\/doi.org\/10.1017\/s089006042400026x","relation":{},"ISSN":["0890-0604","1469-1760"],"issn-type":[{"type":"print","value":"0890-0604"},{"type":"electronic","value":"1469-1760"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"\u00a9 The Author(s), 2024. Published by Cambridge University Press","name":"copyright","label":"Copyright","group":{"name":"copyright_and_licensing","label":"Copyright and Licensing"}}],"article-number":"e22"}}