{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:50:55Z","timestamp":1765547455286,"version":"3.44.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2025,8,30]],"date-time":"2025-08-30T00:00:00Z","timestamp":1756512000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,30]],"date-time":"2025-08-30T00:00:00Z","timestamp":1756512000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100004480","name":"Natural Science Foundation of Shanxi Province","doi-asserted-by":"crossref","award":["202303021211153"],"award-info":[{"award-number":["202303021211153"]}],"id":[{"id":"10.13039\/501100004480","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Shanxi Key Laboratory of Machine Vision and Virtual Reality","award":["447- 110103"],"award-info":[{"award-number":["447- 110103"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62272426"],"award-info":[{"award-number":["62272426"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Shanxi Province Science and Technology Major Special Project","award":["202201150401021"],"award-info":[{"award-number":["202201150401021"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cluster Comput"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s10586-025-05195-2","type":"journal-article","created":{"date-parts":[[2025,8,30]],"date-time":"2025-08-30T10:39:14Z","timestamp":1756550354000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Indoor object detection algorithm based on SBP-YOLOv7"],"prefix":"10.1007","volume":"28","author":[{"given":"Huiyan","family":"Han","sequence":"first","affiliation":[]},{"given":"Wanning","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xie","family":"Han","sequence":"additional","affiliation":[]},{"given":"Liqun","family":"Kuang","sequence":"additional","affiliation":[]},{"given":"Xiaowen","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,30]]},"reference":[{"key":"5195_CR1","doi-asserted-by":"publisher","first-page":"18772","DOI":"10.1109\/ACCESS.2023.3245521","volume":"11","author":"VD Husz\u00e1r","year":"2023","unstructured":"Husz\u00e1r, V.D., Adhikarla, V.K., N\u00e9gyesi, I., et al.: Toward fast and accurate violence detection for automated video surveillance applications. IEEE Access 11, 18772\u201318793 (2023)","journal-title":"IEEE Access"},{"issue":"2","key":"5195_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.asej.2023.102387","volume":"15","author":"AB Atitallah","year":"2024","unstructured":"Atitallah, A.B., Said, Y., Atitallah, M.A.B., et al.: An effective obstacle detection system using deep learning advantages to aid blind and visually impaired navigation. Ain Shams Eng. J. 15(2), 102387 (2024)","journal-title":"Ain Shams Eng. J."},{"key":"5195_CR3","doi-asserted-by":"crossref","unstructured":"Zhao, L., Deng, X., Li, R., Gui, X., Sun, J., Li, T., Zhang, B.: Graph-based robust localization of object-level map for mobile robotic navigation. IEEE Trans. Ind. Electron. 71(1), 697\u2013707 (2023)","DOI":"10.1109\/TIE.2023.3245208"},{"key":"5195_CR4","doi-asserted-by":"publisher","first-page":"1174138","DOI":"10.3389\/fimmu.2023.1174138","volume":"14","author":"P Farhangnia","year":"2023","unstructured":"Farhangnia, P., Ghomi, S.M., Mollazadehghomi, S., et al.: SLAM-family receptors come of age as a potential molecular target in cancer immunotherapy. Front. Immunol. 14, 1174138 (2023)","journal-title":"Front. Immunol."},{"key":"5195_CR5","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vision 60, 91\u2013110 (2004)","journal-title":"Int. J. Comput. Vision"},{"key":"5195_CR6","unstructured":"D\u2019Souza, D., Yampolskiy, R.V.: Baseline Avatar Face Detection using an extended set of Haar-like features. In: Midwest Artificial Intelligence and Cognitive Science Conference, vol. 72 (2012)"},{"issue":"4","key":"5195_CR7","first-page":"1211","volume":"15","author":"RYO Matsumura","year":"2019","unstructured":"Matsumura, R.Y.O., Hanazawa, A.: Human detection using color contrast-based histograms of oriented gradients. Int. J. Innov. Comput., Inf. Control 15(4), 1211\u20131222 (2019)","journal-title":"Int. J. Innov. Comput., Inf. Control"},{"key":"5195_CR8","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., et al.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587. IEEE, Columbus (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"5195_CR9","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440\u20131448. IEEE, Santiago (2015)","DOI":"10.1109\/ICCV.2015.169"},{"issue":"6","key":"5195_CR10","first-page":"1137","volume":"39","author":"SQ Ren","year":"2016","unstructured":"Ren, S.Q., He, K.M., Girshick, R., et al.: Faster R-CNN: towards real-time object detection with region proposal networks. Adv. Neural. Inf. Process. Syst. 39(6), 1137\u20131149 (2016)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5195_CR11","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: Yolov4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)"},{"key":"5195_CR12","unstructured":"Chu, Y.L., Hong, L.J., Kai, H.W.: YOLOv6: a single-stage object detection framework for industrial applications. In: Proceedings of the 2022 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18\u201324 June (2022)"},{"key":"5195_CR13","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the 2022 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18\u201324 June (2022)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"5195_CR14","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., et al.: SSD: single shot multibox detector. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part I 14, pp. 21\u201337. Springer (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"5195_CR15","doi-asserted-by":"publisher","first-page":"1036","DOI":"10.1631\/FITEE.1800275","volume":"20","author":"Q Wang","year":"2019","unstructured":"Wang, Q., Fan, Z., Sheng, W., et al.: Finding misplaced items using a mobile robot in a smart home environment. Front. Inf. Technol. Electron. Eng. 20, 1036\u20131048 (2019)","journal-title":"Front. Inf. Technol. Electron. Eng."},{"key":"5195_CR16","doi-asserted-by":"publisher","first-page":"31645","DOI":"10.1007\/s11042-020-09662-3","volume":"79","author":"M Afif","year":"2020","unstructured":"Afif, M., Ayachi, R., Pissaloux, E., et al.: Indoor objects detection and recognition for ICT mobility assistance of visually impaired people. Multimedia Tools Appl. 79, 31645\u201331662 (2020)","journal-title":"Multimedia Tools Appl."},{"key":"5195_CR17","doi-asserted-by":"crossref","unstructured":"Chen, M., Ren, X., Yan, Z.: Real-time indoor object detection based on deep learning and gradient harmonizing mechanism. In: 2020 IEEE 9th Data Driven Control and Learning Systems Conference (DDCLS), pp. 772\u2013777. IEEE (2020)","DOI":"10.1109\/DDCLS49620.2020.9275060"},{"key":"5195_CR18","doi-asserted-by":"publisher","DOI":"10.3389\/fnbot.2022.881021","volume":"16","author":"L Huang","year":"2022","unstructured":"Huang, L., Chen, C., Yun, J., et al.: Multi-scale feature fusion convolutional neural network for indoor small target detection. Front. Neurorobot. 16, 881021 (2022)","journal-title":"Front. Neurorobot."},{"key":"5195_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.116205","volume":"190","author":"K Hameed","year":"2022","unstructured":"Hameed, K., Chai, D., Rassau, A.: Score-based mask edge improvement of Mask-RCNN for segmentation of fruit and vegetables. Expert Syst. Appl. 190, 116205 (2022)","journal-title":"Expert Syst. Appl."},{"key":"5195_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, D., Zhang, W., Li, F., et al.: Indoor target detection and pose estimation based on point cloud. In: Journal of Physics: Conference Series, vol. 2218, issue 1, p. 012037. IOP Publishing (2022)","DOI":"10.1088\/1742-6596\/2218\/1\/012037"},{"key":"5195_CR21","doi-asserted-by":"crossref","unstructured":"Li, M., Yang, T., Xu, S., et al.: Multi-scale small target detection for indoor mobile rescue vehicles based on improved YOLOv5. In: International Journal of Automotive Technology, pp. 1\u201314 (2024)","DOI":"10.1007\/s12239-024-00104-9"},{"key":"5195_CR22","unstructured":"Yang, L., Zhang, R.Y., Li, L., et al.: SimAM: a simple, parameter-free attention module for convolutional neural networks. In: International Conference on Machine Learning, pp. 11863\u201311874. PMLR (2021)"},{"key":"5195_CR23","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, Q.V.: Efficientdet: scalable and efficient object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10781\u201310790 (2020)","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"5195_CR24","doi-asserted-by":"crossref","unstructured":"Chen, J., Kao, S., He, H., et al.: Run, don't walk: Chasing higher FLOPS for faster neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12021\u201312031 (2023)","DOI":"10.1109\/CVPR52729.2023.01157"},{"key":"5195_CR25","unstructured":"Li, H., Li, J., Wei, H., et al.: Slim-neck by GSConv: a better design paradigm of detector architectures for autonomous vehicles.arXiv preprint arXiv:2206.02424 (2022)"},{"key":"5195_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2023.108562","volume":"217","author":"Y Jia","year":"2024","unstructured":"Jia, Y., Fu, K., Lan, H., et al.: Maize tassel detection with CA-YOLO for UAV images in complex field environments. Comput. Electron. Agric. 217, 108562 (2024)","journal-title":"Comput. Electron. Agric."},{"issue":"1","key":"5195_CR27","doi-asserted-by":"publisher","first-page":"4486","DOI":"10.1038\/s41598-024-54428-8","volume":"14","author":"D Peng","year":"2024","unstructured":"Peng, D., Ding, W., Zhen, T.: A novel low light object detection method based on the YOLOv5 fusion feature enhancement. Sci. Rep. 14(1), 4486 (2024)","journal-title":"Sci. Rep."},{"issue":"1","key":"5195_CR28","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1007\/s11554-023-01387-4","volume":"21","author":"Q Yang","year":"2024","unstructured":"Yang, Q., Meng, H., Gao, Y., et al.: A real-time object detection method for underwater complex environments based on FasterNet-YOLOv7. J. Real-Time Image Proc. 21(1), 8 (2024)","journal-title":"J. Real-Time Image Proc."},{"issue":"8","key":"5195_CR29","doi-asserted-by":"publisher","first-page":"1988","DOI":"10.3390\/agronomy13081988","volume":"13","author":"Z Liang","year":"2023","unstructured":"Liang, Z., Cui, G., Xiong, M., et al.: YOLO-C: an efficient and robust detection algorithm for mature long staple cotton targets with high-resolution RGB images. Agronomy 13(8), 1988 (2023)","journal-title":"Agronomy"},{"key":"5195_CR30","doi-asserted-by":"crossref","unstructured":"Song, Z., Huang, X., Ji, C., et al.: Intelligent identification method of hydrophobic grade of composite insulator based on efficient GA\u2010YOLO former network. IEEJ Trans. Electr. Electron. Eng. 18(7), 1160\u20131175 (2023)","DOI":"10.1002\/tee.23822"},{"key":"5195_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.resconrec.2021.106090","volume":"178","author":"Z Chen","year":"2022","unstructured":"Chen, Z., Yang, J., Chen, L., et al.: Garbage classification system based on improved ShuffleNet v2. Resour. Conserv. Recycl. 178, 106090 (2022)","journal-title":"Resour. Conserv. Recycl."},{"key":"5195_CR32","doi-asserted-by":"crossref","unstructured":"Yu, K., Tang, G., Chen, W., et al.: MobileNet-YOLO v5s: an improved lightweight method for real-time detection of sugarcane stem nodes in complex natural environments. IEEE Access. 11, 104070\u2013104083 (2023)","DOI":"10.1109\/ACCESS.2023.3317951"},{"issue":"14","key":"5195_CR33","doi-asserted-by":"publisher","first-page":"3176","DOI":"10.3390\/electronics12143176","volume":"12","author":"Z Huang","year":"2023","unstructured":"Huang, Z., Wu, J., Su, L., et al.: SP-YOLO-Lite: a lightweight violation detection algorithm based on SP attention mechanism. Electronics 12(14), 3176 (2023)","journal-title":"Electronics"},{"key":"5195_CR34","doi-asserted-by":"publisher","first-page":"5602595","DOI":"10.1155\/2023\/5602595","volume":"2023","author":"G Wang","year":"2023","unstructured":"Wang, G., Feng, A., Gu, C., et al.: YOLO-DFD: a lightweight method for dog feces detection based on improved YOLOv4. J. Sens. 2023, 5602595 (2023)","journal-title":"J. Sens."},{"key":"5195_CR35","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"5195_CR36","doi-asserted-by":"crossref","unstructured":"Wang, Q.L., Wu, B.G., Zhu, P.F., et al.: ECA-Net: efficient channel attention for deep convolutional neural networks. In: Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition, pp. 11531\u201311539 (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"5195_CR37","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., et al.: Cbam: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"5195_CR38","unstructured":"Liu, Y., Shao, Z., Hoffmann, N.: Global attention mechanism: retain information to enhance channel-spatial interactions. arXiv:2112.05561 (2021)"},{"key":"5195_CR39","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., et al.: Microsoft Coco: Common Objects in Context, pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"5195_CR40","doi-asserted-by":"crossref","unstructured":"Fu, Z., Zhang, F., Ren, X., et al.: LE-YOLO: lightweight and efficient detection model for wind turbine blade defects based on improved YOLO. IEEE Access, 12, 135985\u2013135998 (2024)","DOI":"10.1109\/ACCESS.2024.3463391"},{"issue":"2","key":"5195_CR41","doi-asserted-by":"publisher","first-page":"023049","DOI":"10.1117\/1.JEI.33.2.023049","volume":"33","author":"BS Chen","year":"2024","unstructured":"Chen, B.S., Liu, Z.: Lightweight target detection algorithm based on partial convolution. J. Electron. Imaging 33(2), 023049\u2013023049 (2024)","journal-title":"J. Electron. Imaging"},{"issue":"15","key":"5195_CR42","doi-asserted-by":"publisher","first-page":"2076","DOI":"10.3390\/app14156467","volume":"14","author":"Y Zhang","year":"2024","unstructured":"Zhang, Y., Feng, T., Song, Y., et al.: An improved target network model for rail surface defect detection. Appl. Sci. 14(15), 2076\u20133417 (2024)","journal-title":"Appl. Sci."}],"container-title":["Cluster Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05195-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10586-025-05195-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05195-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,17]],"date-time":"2025-09-17T21:22:41Z","timestamp":1758144161000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10586-025-05195-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,30]]},"references-count":42,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["5195"],"URL":"https:\/\/doi.org\/10.1007\/s10586-025-05195-2","relation":{},"ISSN":["1386-7857","1573-7543"],"issn-type":[{"type":"print","value":"1386-7857"},{"type":"electronic","value":"1573-7543"}],"subject":[],"published":{"date-parts":[[2025,8,30]]},"assertion":[{"value":"30 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 January 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 February 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 August 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 September 2025","order":5,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Update","order":6,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The original online version of this article was revised: The affiliation of the authors Wanning Li and Xie Han have been numbered twice to their names. The affiliation numbering has been corrected. The funding note has been corrected from \u2018This research was funded by National Natural Science Foundation of China (No. 62272426) and Natural Science Foundation of Shanxi Province (No. 202303021211153) and Shanxi Province Science and Technology Major Special Project (Grant No. 202201150401021) and Shanxi Key Laboratory of Machine Vision and Virtual Reality (No. 447-110103)\u2019 to \u2018This research was funded by Natural Science Foundation of Shanxi Province (No. 202303021211153) and National Natural Science Foundation of China (No. 62272426) and Shanxi Province Science and Technology Major Special Project (No.  202201150401021) and Shanxi Key Laboratory of Machine Vision and Virtual Reality (No. 447-110103)\u2019.","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. Consent to participate All authors agreed to participate in this paper. The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"558"}}