{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T00:37:01Z","timestamp":1773707821379,"version":"3.50.1"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T00:00:00Z","timestamp":1732320000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T00:00:00Z","timestamp":1732320000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Fujian Province Science and Technology Plan Guided Fund","award":["2021H0013"],"award-info":[{"award-number":["2021H0013"]}]},{"name":"Fujian Province Science and Technology Plan Guided Fund","award":["2021H0013"],"award-info":[{"award-number":["2021H0013"]}]},{"name":"Fujian Province Science and Technology SME Innovation Fund","award":["2021C0019"],"award-info":[{"award-number":["2021C0019"]}]},{"name":"Fujian Province Science and Technology SME Innovation Fund","award":["2021C0019"],"award-info":[{"award-number":["2021C0019"]}]},{"name":"Fujian Province Science and Technology SME Innovation Fund","award":["2021C0019"],"award-info":[{"award-number":["2021C0019"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s11554-024-01573-y","type":"journal-article","created":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T05:53:51Z","timestamp":1732341231000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Intelligent detection of safety helmets and reflective vests based on deep learning"],"prefix":"10.1007","volume":"22","author":[{"given":"Conggong","family":"Lin","sequence":"first","affiliation":[]},{"given":"Yushi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Guodong","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,23]]},"reference":[{"issue":"1","key":"1573_CR1","doi-asserted-by":"publisher","first-page":"73","DOI":"10.3390\/buildings12010073","volume":"12","author":"F Rodrigues","year":"2022","unstructured":"Rodrigues, F., Baptista, J.S., Pinto, D.: Bim approach in construction safety-a case study on preventing falls from height. Buildings 12(1), 73 (2022). https:\/\/doi.org\/10.3390\/buildings12010073","journal-title":"Buildings"},{"key":"1573_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.autcon.2017.09.018","volume":"85","author":"Q Fang","year":"2018","unstructured":"Fang, Q., Li, H., Luo, X., Ding, L., Luo, H., Rose, T.M., et al.: Detecting non-hardhat-use by a deep learning method from far-field surveillance videos. Autom. Constr. 85, 1\u20139 (2018). https:\/\/doi.org\/10.1016\/j.autcon.2017.09.018","journal-title":"Autom. Constr."},{"key":"1573_CR3","doi-asserted-by":"publisher","first-page":"102894","DOI":"10.1016\/j.autcon.2019.102894","volume":"106","author":"J Wu","year":"2019","unstructured":"Wu, J., Cai, N., Chen, W., Wang, H., Wang, G.: Automatic detection of hardhats worn by construction personnel: A deep learning approach and benchmark dataset. Autom. Constr. 106, 102894 (2019). https:\/\/doi.org\/10.1016\/j.autcon.2019.102894","journal-title":"Autom. Constr."},{"issue":"2","key":"1573_CR4","doi-asserted-by":"publisher","first-page":"04018066","DOI":"10.1061\/(ASCE)CP.1943-5487.0000813","volume":"33","author":"BE Mneymneh","year":"2019","unstructured":"Mneymneh, B.E., Abbas, M., Khoury, H.: Vision-based framework for intelligent monitoring of hardhat wearing on construction sites. J. Comput. Civ. Eng. 33(2), 04018066 (2019). https:\/\/doi.org\/10.1061\/(ASCE)CP.1943-5487.0000813","journal-title":"J. Comput. Civ. Eng."},{"key":"1573_CR5","doi-asserted-by":"publisher","unstructured":"Xie, Z., Liu, H., Li, Z., He, Y.: A convolutional neural network based approach towards real-time hard hat detection. In: 2018 IEEE International Conference on Progress in Informatics and Computing (PIC). ieee; p. 430\u2013434 (2018). https:\/\/doi.org\/10.1109\/pic.2018.8706269","DOI":"10.1109\/pic.2018.8706269"},{"key":"1573_CR6","doi-asserted-by":"publisher","unstructured":"Chen, S., Tang, W., Ji, T., Zhu, H., Ouyang, Y., Wang, W.: Detection of safety helmet wearing based on improved faster r-cnn. In: 2020 International joint conference on neural networks (IJCNN). IEEE; p. 1\u20137 (2020). https:\/\/doi.org\/10.1109\/ijcnn48605.2020.9207574","DOI":"10.1109\/ijcnn48605.2020.9207574"},{"key":"1573_CR7","doi-asserted-by":"publisher","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition. p. 779\u2013788 (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.91","DOI":"10.1109\/cvpr.2016.91"},{"key":"1573_CR8","doi-asserted-by":"publisher","unstructured":"Wu, F., Jin, G., Gao, M., Zhiwei, H., Yang, Y.: Helmet detection based on improved yolo v3 deep model. In: 2019 IEEE 16th International conference on networking, sensing and control (ICNSC). IEEE; p. 363\u2013368 (2019b). https:\/\/doi.org\/10.1109\/icnsc.2019.8743246","DOI":"10.1109\/icnsc.2019.8743246"},{"key":"1573_CR9","doi-asserted-by":"publisher","unstructured":"Redmon, J., Farhadi, A.: Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767, https:\/\/doi.org\/10.48550\/arXiv.1804.02767 (2018)","DOI":"10.48550\/arXiv.1804.02767"},{"key":"1573_CR10","doi-asserted-by":"publisher","unstructured":"Huang, G., Liu, Z., Van Der\u00a0Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. p. 4700\u20134708 (2017). https:\/\/doi.org\/10.1109\/cvpr.2017.243","DOI":"10.1109\/cvpr.2017.243"},{"key":"1573_CR11","doi-asserted-by":"publisher","unstructured":"Benyang, D., Xiaochun, L., Miao, Y.: Safety helmet detection method based on yolo v4. In: 2020 16th International conference on computational intelligence and security (CIS). IEEE; p. 155\u2013158 (2020). https:\/\/doi.org\/10.1109\/cis52066.2020.00041","DOI":"10.1109\/cis52066.2020.00041"},{"issue":"1","key":"1573_CR12","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1007\/s11554-023-01268-w","volume":"20","author":"Z Chen","year":"2023","unstructured":"Chen, Z., Zhang, F., Liu, H., Wang, L., Zhang, Q., Guo, L.: Real-time detection algorithm of helmet and reflective vest based on improved yolov5. J. Real-Time Image Proc. 20(1), 4 (2023). https:\/\/doi.org\/10.1007\/s11554-023-01268-w","journal-title":"J. Real-Time Image Proc."},{"key":"1573_CR13","doi-asserted-by":"publisher","DOI":"10.5281\/zenodo.7002879","author":"G Jocher","year":"2022","unstructured":"Jocher, G., Chaurasia, A., Stoken, A., Borovec, J., Kwon, Y., Michael, K., et al.: ultralytics\/yolov5: v6. 2-yolov5 classification models, apple m1, reproducibility, clearml and deci. ai integrations. Zenodo (2022). https:\/\/doi.org\/10.5281\/zenodo.7002879","journal-title":"Zenodo"},{"key":"1573_CR14","doi-asserted-by":"publisher","unstructured":"Milanovic, A., Jovanovic, L., Zivkovic, M., Bacanin, N., Cajic, M., Antonijevic, M.: Exploring pre-trained model potential for reflective vest real time detection with yolov8 models. In: 2024 3rd International Conference on Applied Artificial Intelligence and Computing (ICAAIC). IEEE; p. 1210\u20131216 (2024). https:\/\/doi.org\/10.1109\/icaaic60222.2024.10575617","DOI":"10.1109\/icaaic60222.2024.10575617"},{"key":"1573_CR15","doi-asserted-by":"publisher","unstructured":"Wu, C., Wu, F., Ge, S., Qi, T., Huang, Y., Xie, X.: Neural news recommendation with multi-head self-attention. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). p. 6389\u20136394 (2019c). https:\/\/doi.org\/10.18653\/v1\/d19-1671","DOI":"10.18653\/v1\/d19-1671"},{"key":"1573_CR16","doi-asserted-by":"publisher","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., et\u00a0al.: Attention is all you need. Advances in neural information processing systems 30, (2017) . https:\/\/doi.org\/10.48550\/ARXIV.1706.03762","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"1573_CR17","doi-asserted-by":"publisher","unstructured":"Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition. p. 7263\u20137271 (2017). https:\/\/doi.org\/10.1109\/cvpr.2017.690","DOI":"10.1109\/cvpr.2017.690"},{"key":"1573_CR18","doi-asserted-by":"publisher","unstructured":"Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., et\u00a0al.: Yolov6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976. https:\/\/doi.org\/10.48550\/arXiv.2209.02976 (2022)","DOI":"10.48550\/arXiv.2209.02976"},{"key":"1573_CR19","doi-asserted-by":"publisher","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. p. 7464\u20137475 (2023). https:\/\/doi.org\/10.1109\/cvpr52729.2023.00721","DOI":"10.1109\/cvpr52729.2023.00721"},{"key":"1573_CR20","doi-asserted-by":"publisher","first-page":"122256","DOI":"10.1016\/j.eswa.2023.122256","volume":"238","author":"Y Zhou","year":"2024","unstructured":"Zhou, Y.: A yolo-nl object detector for real-time detection. Expert Syst. Appl. 238, 122256 (2024). https:\/\/doi.org\/10.1016\/j.eswa.2023.122256","journal-title":"Expert Syst. Appl."},{"key":"1573_CR21","doi-asserted-by":"publisher","unstructured":"Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-iou loss: Faster and better learning for bounding box regression. In: Proceedings of the AAAI conference on artificial intelligence; vol.\u00a034. p. 12993\u201313000 (2020). https:\/\/doi.org\/10.1609\/aaai.v34i07.6999","DOI":"10.1609\/aaai.v34i07.6999"},{"key":"1573_CR22","doi-asserted-by":"publisher","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S.: Generalized intersection over union: A metric and a loss for bounding box regression. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. p. 658\u2013666 (2019). https:\/\/doi.org\/10.1109\/cvpr.2019.00075","DOI":"10.1109\/cvpr.2019.00075"},{"key":"1573_CR23","doi-asserted-by":"publisher","unstructured":"Bolya, D., Foley, S., Hays, J., Hoffman, J.: Tide: A general toolbox for identifying object detection errors. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part III 16. Springer; p. 558\u2013573 (2020). https:\/\/doi.org\/10.1007\/978-3-030-58580-8_33","DOI":"10.1007\/978-3-030-58580-8_33"},{"key":"1573_CR24","doi-asserted-by":"publisher","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision. p. 2980\u20132988 (2017). https:\/\/doi.org\/10.1109\/iccv.2017.324","DOI":"10.1109\/iccv.2017.324"},{"key":"1573_CR25","doi-asserted-by":"publisher","unstructured":"Zhang, S., Chi, C., Yao, Y., Lei, Z., Li, S.Z.: Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. p. 9759\u20139768 (2020). https:\/\/doi.org\/10.1109\/cvpr42600.2020.00978","DOI":"10.1109\/cvpr42600.2020.00978"},{"key":"1573_CR26","doi-asserted-by":"publisher","unstructured":"Liu, S., Zeng, Z., Ren, T., Li, F., Zhang, H., Yang, J., et\u00a0al.: Grounding dino: Marrying dino with grounded pre-training for open-set object detection. arXiv preprint arXiv:2303.05499, https:\/\/doi.org\/10.48550\/arXiv.2303.05499 (2023)","DOI":"10.48550\/arXiv.2303.05499"},{"key":"1573_CR27","doi-asserted-by":"publisher","unstructured":"Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: Yolox: Exceeding yolo series in 2021. arXiv preprint arXiv:2107.08430. https:\/\/doi.org\/10.48550\/arXiv.2107.08430 (2021)","DOI":"10.48550\/arXiv.2107.08430"},{"key":"1573_CR28","doi-asserted-by":"publisher","unstructured":"Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV). p. 3\u201319 (2018). https:\/\/doi.org\/10.48550\/arXiv.1807.06521","DOI":"10.48550\/arXiv.1807.06521"},{"key":"1573_CR29","doi-asserted-by":"publisher","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. p. 7132\u20137141 (2018). https:\/\/doi.org\/10.1109\/cvpr.2018.00745","DOI":"10.1109\/cvpr.2018.00745"},{"key":"1573_CR30","doi-asserted-by":"publisher","unstructured":"Ouyang, D., He, S., Zhang, G., Luo, M., Guo, H., Zhan, J., et\u00a0al.: Efficient multi-scale attention module with cross-spatial learning. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; p. 1\u20135 (2023). https:\/\/doi.org\/10.1109\/icassp49357.2023.10096516","DOI":"10.1109\/icassp49357.2023.10096516"},{"key":"1573_CR31","doi-asserted-by":"publisher","first-page":"146","DOI":"10.1016\/j.neucom.2022.07.042","volume":"506","author":"YF Zhang","year":"2022","unstructured":"Zhang, Y.F., Ren, W., Zhang, Z., Jia, Z., Wang, L., Tan, T.: Focal and efficient iou loss for accurate bounding box regression. Neurocomputing 506, 146\u2013157 (2022). https:\/\/doi.org\/10.1016\/j.neucom.2022.07.042","journal-title":"Neurocomputing"},{"key":"1573_CR32","doi-asserted-by":"publisher","unstructured":"Gevorgyan, Z.: Siou loss: More powerful learning for bounding box regression. arXiv preprint arXiv:2205.12740, https:\/\/doi.org\/10.48550\/arXiv.2205.12740 (2022)","DOI":"10.48550\/arXiv.2205.12740"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01573-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-024-01573-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01573-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,3]],"date-time":"2025-02-03T17:18:32Z","timestamp":1738603112000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-024-01573-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,23]]},"references-count":32,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["1573"],"URL":"https:\/\/doi.org\/10.1007\/s11554-024-01573-y","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,23]]},"assertion":[{"value":"27 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 October 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 November 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"None.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"5"}}