{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T18:01:22Z","timestamp":1764784882978,"version":"build-2065373602"},"reference-count":34,"publisher":"MDPI AG","issue":"5","license":[{"start":{"date-parts":[[2022,4,26]],"date-time":"2022-04-26T00:00:00Z","timestamp":1650931200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Symmetry"],"abstract":"<jats:p>One of the most noticeable characteristics of security issues is the prevalence of \u201cSecurity Asymmetry\u201d. The safety of production and even the lives of workers can be jeopardized if risk factors aren\u2019t detected in time. Today, object detection technology plays a vital role in actual operating conditions. For the sake of warning danger and ensuring the work security, we propose the Attention-guided Feature Fusion Network method and apply it to the Helmet Detection in this paper. AFFN method, which is capable of reliably detecting objects of a wider range of sizes, outperforms previous methods with an mAP value of 85.3% and achieves an excellent result in helmet detection with an mAP value of 62.4%. From objects of finite sizes to a wider range of sizes, the proposed method achieves \u201csymmetry\u201d in the sense of detection.<\/jats:p>","DOI":"10.3390\/sym14050887","type":"journal-article","created":{"date-parts":[[2022,4,26]],"date-time":"2022-04-26T21:37:53Z","timestamp":1651009073000},"page":"887","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Object Detection by Attention-Guided Feature Fusion Network"],"prefix":"10.3390","volume":"14","author":[{"given":"Yuxuan","family":"Shi","sequence":"first","affiliation":[{"name":"School of Statistics and Data Science, Nankai University, No. 94 Weijin Road, Nankai District, Tianjin 300071, China"}]},{"given":"Yue","family":"Fan","sequence":"additional","affiliation":[{"name":"School of Journalism and Communication, Nankai University, No. 94 Weijin Road, Nankai District, Tianjin 300071, China"}]},{"given":"Siqi","family":"Xu","sequence":"additional","affiliation":[{"name":"School of Journalism and Communication, Nankai University, No. 94 Weijin Road, Nankai District, Tianjin 300071, China"}]},{"given":"Yue","family":"Gao","sequence":"additional","affiliation":[{"name":"Zhou Enlai School of Government, Nankai University, No. 94 Weijin Road, Nankai District, Tianjin 300071, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3002-0662","authenticated-orcid":false,"given":"Ran","family":"Gao","sequence":"additional","affiliation":[{"name":"School of Journalism and Communication, Nankai University, No. 94 Weijin Road, Nankai District, Tianjin 300071, China"}]}],"member":"1968","published-online":{"date-parts":[[2022,4,26]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"261","DOI":"10.1007\/s11263-019-01247-4","article-title":"Deep learning for generic object detection: A survey","volume":"128","author":"Liu","year":"2020","journal-title":"Int. J. Comput. Vis."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27\u201330). You only look once: Unified, real-time object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Redmon, J., and Farhadi, A. (2017, January 21\u201326). YOLO9000: Better, faster, stronger. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.690"},{"key":"ref_4","unstructured":"Redmon, J., and Farhadi, A. (2018). Yolov3: An incremental improvement. arXiv."},{"key":"ref_5","first-page":"1","article-title":"Faster r-cnn: Towards real-time object detection with region proposal networks","volume":"28","author":"Ren","year":"2015","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2017, January 21\u201326). Feature pyramid networks for object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.106"},{"key":"ref_8","first-page":"1","article-title":"Poly-YOLO: Higher speed, more precise detection and instance segmentation for YOLOv3","volume":"27","author":"Hurtik","year":"2022","journal-title":"Neural Comput. Appl."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"33","DOI":"10.47933\/ijeir.1027304","article-title":"A Factorial Analysis of Industrial Safety","volume":"4","year":"2022","journal-title":"Int. J. Eng. Innov. Res."},{"key":"ref_10","unstructured":"Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"142","DOI":"10.1109\/TPAMI.2015.2437384","article-title":"Region-based convolutional networks for accurate object detection and segmentation","volume":"38","author":"Girshick","year":"2015","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015, January 7\u201313). Fast r-cnn. Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile.","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Zhu, X., Wang, Y., Dai, J., Yuan, L., and Wei, Y. (2017, January 22\u201329). Flow-guided feature aggregation for video object detection. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.52"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Liu, J., Zhang, W., Tang, Y., Tang, J., and Wu, G. (2020, January 14\u201319). Residual feature aggregation network for image super-resolution. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00243"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Li, H., Xiong, P., Fan, H., and Sun, J. (2019, January 15\u201320). Dfanet: Deep feature aggregation for real-time semantic segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00975"},{"key":"ref_16","first-page":"1","article-title":"MobileSal: Extremely efficient RGB-D salient object detection","volume":"12","author":"Wu","year":"2021","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_17","first-page":"1","article-title":"Imagenet classification with deep convolutional neural networks","volume":"25","author":"Krizhevsky","year":"2012","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., and Sun, G. (2018, January 18\u201322). Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"1949","DOI":"10.1109\/TIP.2021.3049959","article-title":"Bilateral attention network for RGB-D salient object detection","volume":"30","author":"Zhang","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., and Kweon, I.S. (2018, January 8\u201314). Cbam: Convolutional block attention module. Proceedings of the European conference on computer vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_21","unstructured":"Park, J., Woo, S., Lee, J.Y., and Kweon, I.S. (2018). Bam: Bottleneck attention module. arXiv."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Niu, B., Wen, W., Ren, W., Zhang, X., Yang, L., Wang, S., Zhang, K., Cao, X., and Shen, H. (2020). Single image super-resolution via a holistic attention network. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-030-58610-2_12"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., and Balasubramanian, V.N. (2018, January 12\u201315). Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), IEEE, Lake Tahoe, NV, USA.","DOI":"10.1109\/WACV.2018.00097"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhang, L., Cheng, M.M., and Feng, J. (2020, January 14\u201319). Strip pooling: Rethinking spatial pooling for scene parsing. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00406"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Xu, G., Xu, J., Li, Z., Wang, L., Sun, X., and Cheng, M.M. (2021, January 19\u201325). Temporal modulation network for controllable space-time video super-resolution. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Virtual.","DOI":"10.1109\/CVPR46437.2021.00632"},{"key":"ref_26","unstructured":"Yang, L., Zhang, R.Y., Li, L., and Xie, X. (2021). Simam: A simple, parameter-free attention module for convolutional neural networks. International Conference on Machine Learning, PMLR."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","article-title":"The pascal visual object classes (voc) challenge","volume":"88","author":"Everingham","year":"2010","journal-title":"Int. J. Comput. Vis."},{"key":"ref_28","first-page":"1","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume":"32","author":"Paszke","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_29","unstructured":"Kingma, D.P., and Ba, J. (2014). Adam: A method for stochastic optimization. arXiv."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016). Ssd: Single shot multibox detector. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_31","unstructured":"Fu, C.Y., Liu, W., Ranga, A., Tyagi, A., and Berg, A.C. (2017). Dssd: Deconvolutional single shot detector. arXiv."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Zhou, P., Ni, B., Geng, C., Hu, J., and Xu, Y. (2018, January 18\u201322). Scale-transferrable object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00062"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Zhao, C., Wang, J., Zhao, X., Wu, Y., and Lu, H. (2017, January 22\u201329). Couplenet: Coupling global structure with local parts for object detection. Proceedings of the IEEE International Conference on COMPUTER Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.444"},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Liu, S., Huang, D., and Wang, Y. (2018, January 8\u201314). Receptive field block net for accurate and fast object detection. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01252-6_24"}],"container-title":["Symmetry"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2073-8994\/14\/5\/887\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T23:01:11Z","timestamp":1760137271000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2073-8994\/14\/5\/887"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,26]]},"references-count":34,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2022,5]]}},"alternative-id":["sym14050887"],"URL":"https:\/\/doi.org\/10.3390\/sym14050887","relation":{},"ISSN":["2073-8994"],"issn-type":[{"type":"electronic","value":"2073-8994"}],"subject":[],"published":{"date-parts":[[2022,4,26]]}}}