{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,5]],"date-time":"2026-04-05T20:34:01Z","timestamp":1775421241801,"version":"3.50.1"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"17","license":[{"start":{"date-parts":[[2023,11,4]],"date-time":"2023-11-04T00:00:00Z","timestamp":1699056000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,11,4]],"date-time":"2023-11-04T00:00:00Z","timestamp":1699056000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Tertiary Education Trust Fund 2020","award":["TETF\/ES\/DR&D-CE\/NRF2020\/SETI\/88\/VOL.1"],"award-info":[{"award-number":["TETF\/ES\/DR&D-CE\/NRF2020\/SETI\/88\/VOL.1"]}]},{"name":"ERASMUS+ KA107"},{"DOI":"10.13039\/501100000780","name":"European Commission","doi-asserted-by":"publisher","award":["H2020-MSCA-RISE-2020\/101008297"],"award-info":[{"award-number":["H2020-MSCA-RISE-2020\/101008297"]}],"id":[{"id":"10.13039\/501100000780","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011033","name":"Agencia Estatal de Investigaci\u00f3n","doi-asserted-by":"publisher","award":["PID2020-115339RB-I00"],"award-info":[{"award-number":["PID2020-115339RB-I00"]}],"id":[{"id":"10.13039\/501100011033","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011033","name":"Agencia Estatal de Investigaci\u00f3n","doi-asserted-by":"publisher","award":["CPP2021-008880"],"award-info":[{"award-number":["CPP2021-008880"]}],"id":[{"id":"10.13039\/501100011033","id-type":"DOI","asserted-by":"publisher"}]},{"name":"DAAD In-country\/In-region PhD scholarship"},{"DOI":"10.13039\/501100007515","name":"Universidad de Valladolid","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100007515","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Deep learning-based object detection models have become a preferred choice for crop detection tasks in crop monitoring activities due to their high accuracy and generalization capabilities. However, their high computational demand and large memory footprint pose a challenge for use on mobile embedded devices deployed in crop monitoring settings. Various approaches have been taken to minimize the computational cost and reduce the size of object detection models such as channel and layer pruning, detection head searching, backbone optimization, etc. In this work, we approached computational lightening, model compression, and speed improvement by discarding one or more of the three detection scales of the YOLOv5 object detection model. Thus, we derived up to five separate fast and light models, each with only one or two detection scales. To evaluate the new models for a real crop monitoring use case, the models were deployed on NVIDIA Jetson nano and NVIDIA Jetson Orin devices. The new models achieved up to 21.4% reduction in giga floating-point operations per second (GFLOPS), 31.9% reduction in number of parameters, 30.8% reduction in model size, 28.1% increase in inference speed, with only a small average accuracy drop of 3.6%. These new models are suitable for crop detection tasks since the crops are usually of similar sizes due to the high likelihood of being in the same growth stage, thus, making it sufficient to detect the crops with just one or two detection scales.<\/jats:p>","DOI":"10.1007\/s11042-023-17435-x","type":"journal-article","created":{"date-parts":[[2023,11,4]],"date-time":"2023-11-04T07:02:53Z","timestamp":1699081373000},"page":"50197-50223","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Simplifying YOLOv5 for deployment in a real crop monitoring setting"],"prefix":"10.1007","volume":"83","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0965-0937","authenticated-orcid":false,"given":"Emmanuel C.","family":"Nnadozie","sequence":"first","affiliation":[]},{"given":"Pablo","family":"Casaseca-de-la-Higuera","sequence":"additional","affiliation":[]},{"given":"Ogechukwu","family":"Iloanusi","sequence":"additional","affiliation":[]},{"given":"Ozoemena","family":"Ani","sequence":"additional","affiliation":[]},{"given":"Carlos","family":"Alberola-L\u00f3pez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,4]]},"reference":[{"key":"17435_CR1","doi-asserted-by":"publisher","unstructured":"Mustafa MM, Hussain A, Ghazali KH, Riyadi S (2007) Implementation of image processing technique in real time vision system for automatic weeding strategy. In: 2007 IEEE international symposium on signal processing and information technology.\u00a0 IEEE, pp 632\u2013635. https:\/\/doi.org\/10.1109\/ISSPIT.2007.4458197","DOI":"10.1109\/ISSPIT.2007.4458197"},{"key":"17435_CR2","doi-asserted-by":"publisher","first-page":"2275","DOI":"10.1016\/j.eswa.2012.10.033","volume":"40","author":"J Romeo","year":"2013","unstructured":"Romeo J, Pajares G, Montalvo M et al (2013) A new expert system for greenness identification in agricultural images. Expert Syst Appl 40:2275\u20132286. https:\/\/doi.org\/10.1016\/j.eswa.2012.10.033","journal-title":"Expert Syst Appl"},{"key":"17435_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1111\/j.1365-3180.2010.00829.x","volume":"51","author":"F L\u00f3pez-Granados","year":"2011","unstructured":"L\u00f3pez-Granados F (2011) Weed detection for site-specific weed management: mapping and real-time approaches. Weed Res 51:1\u201311. https:\/\/doi.org\/10.1111\/j.1365-3180.2010.00829.x","journal-title":"Weed Res"},{"key":"17435_CR4","doi-asserted-by":"publisher","first-page":"312","DOI":"10.1017\/S0021859618000436","volume":"156","author":"A Kamilaris","year":"2018","unstructured":"Kamilaris A, Prenafeta-Bold\u00fa FX (2018) A review of the use of convolutional neural networks in agriculture. J Agric Sci 156:312\u2013322. https:\/\/doi.org\/10.1017\/S0021859618000436","journal-title":"J Agric Sci"},{"key":"17435_CR5","doi-asserted-by":"publisher","unstructured":"Moazzam SI, Khan US, Tiwana MI, et al (2019) A review of application of deep learning for weeds and crops classification in agriculture. In: 2019 international conference on robotics and automation in industry (ICRAI) Rawalpindi, Pakistan, pp 1\u20136. https:\/\/doi.org\/10.1109\/ICRAI47710.2019.8967350","DOI":"10.1109\/ICRAI47710.2019.8967350"},{"key":"17435_CR6","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1016\/j.compag.2017.01.001","volume":"135","author":"J Tang","year":"2017","unstructured":"Tang J, Wang D, Zhang Z et al (2017) Weed identification based on K-means feature learning combined with convolutional neural network. Comput Electron Agric 135:63\u201370. https:\/\/doi.org\/10.1016\/j.compag.2017.01.001","journal-title":"Comput Electron Agric"},{"key":"17435_CR7","doi-asserted-by":"publisher","unstructured":"Czymmek V, Harders LO, Knoll FJ, Hussmann S (2019) Vision-based deep learning approach for real-time detection of weeds in organic farming. In: 2019 IEEE international instrumentation and measurement technology conference (I2MTC), Auckland, New Zealand, pp 1\u20135. https:\/\/doi.org\/10.1109\/I2MTC.2019.8826921","DOI":"10.1109\/I2MTC.2019.8826921"},{"key":"17435_CR8","doi-asserted-by":"publisher","first-page":"3203","DOI":"10.1109\/TCSVT.2021.3095970","volume":"32","author":"H Kirchhoffer","year":"2022","unstructured":"Kirchhoffer H, Haase P, Samek W et al (2022) Overview of the neural network compression and representation (NNR) standard. IEEE Trans Circuits Syst Video Technol 32:3203\u20133216. https:\/\/doi.org\/10.1109\/TCSVT.2021.3095970","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"17435_CR9","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1109\/JETCAS.2021.3129415","volume":"11","author":"AN Mazumder","year":"2021","unstructured":"Mazumder AN, Meng J, Rashid H-A et al (2021) A Survey on the optimization of neural network accelerators for micro-AI on-device inference. IEEE J Emerg Sel Top Circuits Syst 11:532\u2013547. https:\/\/doi.org\/10.1109\/JETCAS.2021.3129415","journal-title":"IEEE J Emerg Sel Top Circuits Syst"},{"key":"17435_CR10","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1109\/JSTSP.2020.3038227","volume":"15","author":"W Sun","year":"2021","unstructured":"Sun W, Chen S, Huang L et al (2021) Deep convolutional neural network compression via coupled tensor decomposition. IEEE J Sel Top Signal Process 15:603\u2013616. https:\/\/doi.org\/10.1109\/JSTSP.2020.3038227","journal-title":"IEEE J Sel Top Signal Process"},{"key":"17435_CR11","doi-asserted-by":"publisher","first-page":"1031","DOI":"10.1093\/imaiai\/iaaa042","volume":"10","author":"S Oymak","year":"2021","unstructured":"Oymak S, Soltanolkotabi M (2021) Learning a deep convolutional neural network via tensor decomposition. Inf Inference 10:1031\u20131071. https:\/\/doi.org\/10.1093\/imaiai\/iaaa042","journal-title":"Inf Inference"},{"key":"17435_CR12","doi-asserted-by":"publisher","first-page":"127917","DOI":"10.1109\/ACCESS.2021.3112695","volume":"9","author":"G Wu","year":"2021","unstructured":"Wu G, Wang S, Liu L (2021) Fast video summary generation based on low rank tensor decomposition. IEEE Access 9:127917\u2013127926. https:\/\/doi.org\/10.1109\/ACCESS.2021.3112695","journal-title":"IEEE Access"},{"key":"17435_CR13","doi-asserted-by":"publisher","first-page":"350","DOI":"10.1016\/j.neunet.2022.02.024","volume":"150","author":"A Nekooei","year":"2022","unstructured":"Nekooei A, Safari S (2022) Compression of deep neural networks based on quantized tensor decomposition to implement on reconfigurable hardware platforms. Neural Netw 150:350\u2013363. https:\/\/doi.org\/10.1016\/j.neunet.2022.02.024","journal-title":"Neural Netw"},{"key":"17435_CR14","doi-asserted-by":"publisher","unstructured":"Qi Q, Lu Y, Li J et al (2021) Learning low resource consumption cnn through pruning and quantization. IEEE Trans Emerg Top Comput 1\u20131. https:\/\/doi.org\/10.1109\/TETC.2021.3050770","DOI":"10.1109\/TETC.2021.3050770"},{"key":"17435_CR15","doi-asserted-by":"publisher","first-page":"6488","DOI":"10.1109\/TCSVT.2022.3167951","volume":"32","author":"E Camci","year":"2022","unstructured":"Camci E, Gupta M, Wu M, Lin J (2022) QLP: Deep Q-learning for pruning deep neural networks. IEEE Trans Circuits Syst Video Technol 32:6488\u20136501. https:\/\/doi.org\/10.1109\/TCSVT.2022.3167951","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"17435_CR16","doi-asserted-by":"publisher","unstructured":"Knight A, Lee BK (2020) Performance analysis of network pruning for deep learning based age-gender estimation. In: 2020 International conference on computational science and computational intelligence (CSCI), Las Vegas, NV, USA, pp. 1684\u20131687. https:\/\/doi.org\/10.1109\/CSCI51800.2020.00310","DOI":"10.1109\/CSCI51800.2020.00310"},{"key":"17435_CR17","doi-asserted-by":"publisher","first-page":"5703","DOI":"10.1109\/TVT.2020.2983143","volume":"69","author":"Y Lin","year":"2020","unstructured":"Lin Y, Tu Y, Dou Z (2020) An improved neural network pruning technology for automatic modulation classification in edge devices. IEEE Trans Veh Technol 69:5703\u20135706. https:\/\/doi.org\/10.1109\/TVT.2020.2983143","journal-title":"IEEE Trans Veh Technol"},{"issue":"241","key":"17435_CR18","first-page":"1","volume":"22","author":"T Hoefler","year":"2021","unstructured":"Hoefler T, Alistarh D, Ben-Nun T et al (2021) Sparsity in deep learning: pruning and growth for efficient inference and training in neural networks. J Mach Learn Res 22(241):1\u2013124.\u00a0http:\/\/jmlr.org\/papers\/v22\/21-0366.html","journal-title":"J Mach Learn Res"},{"key":"17435_CR19","doi-asserted-by":"publisher","first-page":"62605","DOI":"10.1109\/ACCESS.2023.3287975","volume":"11","author":"VA Boateng","year":"2023","unstructured":"Boateng VA, Yang B (2023) A global modeling pruning ensemble stacking with deep learning and neural network meta-learner for passenger train delay prediction. IEEE Access 11:62605\u201362615. https:\/\/doi.org\/10.1109\/ACCESS.2023.3287975","journal-title":"IEEE Access"},{"key":"17435_CR20","doi-asserted-by":"publisher","first-page":"46441","DOI":"10.1109\/ACCESS.2023.3273952","volume":"11","author":"J Li","year":"2023","unstructured":"Li J, Chen X, Zheng P et al (2023) Deep generative knowledge distillation by likelihood finetuning. IEEE Access 11:46441\u201346453. https:\/\/doi.org\/10.1109\/ACCESS.2023.3273952","journal-title":"IEEE Access"},{"key":"17435_CR21","doi-asserted-by":"publisher","first-page":"6985","DOI":"10.1109\/TIP.2021.3101158","volume":"30","author":"Z Feng","year":"2021","unstructured":"Feng Z, Lai J, Xie X (2021) Resolution-aware knowledge distillation for efficient inference. IEEE Trans Image Process 30:6985\u20136996. https:\/\/doi.org\/10.1109\/TIP.2021.3101158","journal-title":"IEEE Trans Image Process"},{"key":"17435_CR22","doi-asserted-by":"publisher","first-page":"1733","DOI":"10.1109\/TCC.2022.3160129","volume":"11","author":"Z Tao","year":"2023","unstructured":"Tao Z, Xia Q, Cheng S, Li Q (2023) An efficient and robust cloud-based deep learning with knowledge distillation. IEEE Trans Cloud Comput 11:1733\u20131745. https:\/\/doi.org\/10.1109\/TCC.2022.3160129","journal-title":"IEEE Trans Cloud Comput"},{"key":"17435_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.105560","volume":"117","author":"M Sepahvand","year":"2023","unstructured":"Sepahvand M, Abdali-Mohammadi F, Taherkordi A (2023) An adaptive teacher\u2013student learning algorithm with decomposed knowledge distillation for on-edge intelligence. Eng Appl Artif Intell 117:105560. https:\/\/doi.org\/10.1016\/j.engappai.2022.105560","journal-title":"Eng Appl Artif Intell"},{"key":"17435_CR24","doi-asserted-by":"publisher","unstructured":"Zoph B, Le QV (2017) Neural architecture search with reinforcement learning. https:\/\/doi.org\/10.48550\/arXiv.1611.01578","DOI":"10.48550\/arXiv.1611.01578"},{"key":"17435_CR25","doi-asserted-by":"publisher","first-page":"25217","DOI":"10.1109\/ACCESS.2023.3253818","volume":"11","author":"KT Chitty-Venkata","year":"2023","unstructured":"Chitty-Venkata KT, Emani M, Vishwanath V, Somani AK (2023) Neural architecture search benchmarks: insights and survey. IEEE Access 11:25217\u201325236. https:\/\/doi.org\/10.1109\/ACCESS.2023.3253818","journal-title":"IEEE Access"},{"key":"17435_CR26","doi-asserted-by":"publisher","first-page":"105890","DOI":"10.1016\/j.engappai.2023.105890","volume":"120","author":"JB Thomas","year":"2023","unstructured":"Thomas JB, Shihabudheen KV (2023) Neural architecture search algorithm to optimize deep Transformer model for fault detection in electrical power distribution systems. Eng Appl Artif Intell 120:105890. https:\/\/doi.org\/10.1016\/j.engappai.2023.105890","journal-title":"Eng Appl Artif Intell"},{"key":"17435_CR27","doi-asserted-by":"publisher","first-page":"108867","DOI":"10.1016\/j.compeleceng.2023.108867","volume":"110","author":"S Khan","year":"2023","unstructured":"Khan S, Rizwan A, Khan AN et al (2023) A multi-perspective revisit to the optimization methods of Neural Architecture Search and Hyper-parameter optimization for non-federated and federated learning environments. Comput Electr Eng 110:108867. https:\/\/doi.org\/10.1016\/j.compeleceng.2023.108867","journal-title":"Comput Electr Eng"},{"key":"17435_CR28","doi-asserted-by":"publisher","first-page":"945","DOI":"10.3390\/electronics11060945","volume":"11","author":"D Ghimire","year":"2022","unstructured":"Ghimire D, Kil D, Kim S (2022) A survey on efficient convolutional neural networks and hardware acceleration. Electronics (Basel) 11:945. https:\/\/doi.org\/10.3390\/electronics11060945","journal-title":"Electronics (Basel)"},{"key":"17435_CR29","doi-asserted-by":"publisher","first-page":"3777","DOI":"10.3390\/s23073777","volume":"23","author":"K Choi","year":"2023","unstructured":"Choi K, Wi SM, Jung HG, Suhr JK (2023) Simplification of deep neural network-based object detector for real-time edge computing. Sensors 23:3777. https:\/\/doi.org\/10.3390\/s23073777","journal-title":"Sensors"},{"key":"17435_CR30","doi-asserted-by":"publisher","unstructured":"Neill JO (2020) An overview of neural network compression. https:\/\/doi.org\/10.48550\/arXiv.2006.03669","DOI":"10.48550\/arXiv.2006.03669"},{"key":"17435_CR31","doi-asserted-by":"publisher","first-page":"466","DOI":"10.1016\/j.neucom.2022.11.029","volume":"518","author":"ES Jeon","year":"2023","unstructured":"Jeon ES, Choi H, Shukla A, Turaga P (2023) Leveraging angular distributions for improved knowledge distillation. Neurocomputing 518:466\u2013481. https:\/\/doi.org\/10.1016\/j.neucom.2022.11.029","journal-title":"Neurocomputing"},{"key":"17435_CR32","doi-asserted-by":"publisher","unstructured":"Kim T, Oh J, Kim N et al (2021) Comparing kullback-leibler divergence and mean squared error loss in knowledge distillation. In: Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, Montreal, pp 2628\u20132635. https:\/\/doi.org\/10.24963\/ijcai.2021\/362","DOI":"10.24963\/ijcai.2021\/362"},{"key":"17435_CR33","doi-asserted-by":"publisher","first-page":"5004","DOI":"10.1109\/TNNLS.2021.3067028","volume":"33","author":"Z Ding","year":"2022","unstructured":"Ding Z, Chen Y, Li N et al (2022) BNAS: efficient neural architecture search using broad scalable architecture. IEEE Trans Neural Netw Learn Syst 33:5004\u20135018. https:\/\/doi.org\/10.1109\/TNNLS.2021.3067028","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"17435_CR34","doi-asserted-by":"publisher","unstructured":"Redmon J, Farhadi A (2018) YOLOv3: an incremental improvement. https:\/\/doi.org\/10.48550\/arXiv.1804.02767","DOI":"10.48550\/arXiv.1804.02767"},{"key":"17435_CR35","doi-asserted-by":"publisher","unstructured":"Girshick R (2015) Fast R-CNN. In: 2015 IEEE international conference on computer vision (ICCV), Santiago, pp 1440\u20131448. https:\/\/doi.org\/10.1109\/ICCV.2015.169","DOI":"10.1109\/ICCV.2015.169"},{"key":"17435_CR36","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2017","unstructured":"Ren S, He K, Girshick R, Sun J (2017) Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39:1137\u20131149. https:\/\/doi.org\/10.1109\/TPAMI.2016.2577031","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"17435_CR37","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1109\/TPAMI.2018.2844175","volume":"42","author":"K He","year":"2020","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2020) Mask R-CNN. IEEE Trans Pattern Anal Mach Intell 42:386\u2013397. https:\/\/doi.org\/10.1109\/TPAMI.2018.2844175","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"17435_CR38","doi-asserted-by":"publisher","unstructured":"Jocher G, Stoken A, Borovec J et al (2020) YOLOv5. https:\/\/doi.org\/10.5281\/ZENODO.4154370","DOI":"10.5281\/ZENODO.4154370"},{"key":"17435_CR39","doi-asserted-by":"publisher","unstructured":"Redmon J, Farhadi A (2017) YOLO9000: better, faster, stronger. Proceedings - 30th IEEE Conference on computer vision and pattern recognition, CVPR 2017 2017-Janua: 6517\u20136525. https:\/\/doi.org\/10.1109\/CVPR.2017.690","DOI":"10.1109\/CVPR.2017.690"},{"key":"17435_CR40","doi-asserted-by":"publisher","unstructured":"Bochkovskiy A, Wang C-Y, Liao H-YM (2020) YOLOv4: Optimal speed and accuracy of object detection. https:\/\/doi.org\/10.48550\/arXiv.2004.10934","DOI":"10.48550\/arXiv.2004.10934"},{"key":"17435_CR41","doi-asserted-by":"publisher","unstructured":"Liu W, Anguelov D, Erhan D et al (2016) SSD: Single shot multibox detector. Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics) 9905 LNCS:21\u201337. https:\/\/doi.org\/10.1007\/978-3-319-46448-0_2\/FIGURES\/5","DOI":"10.1007\/978-3-319-46448-0_2\/FIGURES\/5"},{"key":"17435_CR42","doi-asserted-by":"publisher","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. Proceedings of the IEEE computer society conference on computer vision and pattern recognition 2016-Decem:779\u2013788. https:\/\/doi.org\/10.1109\/CVPR.2016.91","DOI":"10.1109\/CVPR.2016.91"},{"key":"17435_CR43","doi-asserted-by":"publisher","unstructured":"Wang CY, Bochkovskiy A, Liao HYM (2021) Scaled-YOLOv4: scaling cross stage partial network. In: 2021 IEEE\/CVF conference on computer vision and pattern recognition (CVPR). IEEE, pp 13024\u201313033. https:\/\/doi.org\/10.1109\/CVPR46437.2021.01283","DOI":"10.1109\/CVPR46437.2021.01283"},{"key":"17435_CR44","doi-asserted-by":"publisher","unstructured":"Ge Z, Liu S, Wang F et al (2021) YOLOX: exceeding YOLO series in 2021. arXiv. https:\/\/doi.org\/10.48550\/arXiv.2107.08430","DOI":"10.48550\/arXiv.2107.08430"},{"key":"17435_CR45","doi-asserted-by":"publisher","unstructured":"Wang CY, Yeh IH, Liao HYM (2021) You only learn one representation: unified network for multiple tasks. arXiv. https:\/\/doi.org\/10.48550\/arXiv.2105.04206","DOI":"10.48550\/arXiv.2105.04206"},{"key":"17435_CR46","doi-asserted-by":"publisher","unstructured":"Xu S, Wang X, Lv W et al (2022) PP-YOLOE: An evolved version of YOLO. arXiv. https:\/\/doi.org\/10.48550\/arXiv.2203.16250","DOI":"10.48550\/arXiv.2203.16250"},{"key":"17435_CR47","doi-asserted-by":"publisher","unstructured":"Li C, Li L, Jiang H et al (2022) YOLOv6: a single-stage object detection framework for industrial applications. arXiv. https:\/\/doi.org\/10.48550\/arXiv.2209.02976","DOI":"10.48550\/arXiv.2209.02976"},{"key":"17435_CR48","doi-asserted-by":"publisher","unstructured":"Wang CY, Bochkovskiy A, Liao HYM (2022) YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv. https:\/\/doi.org\/10.48550\/arXiv.2207.02696","DOI":"10.48550\/arXiv.2207.02696"},{"key":"17435_CR49","doi-asserted-by":"publisher","unstructured":"Xu X, Jiang Y, Chen W et al (2022) DAMO-YOLO: a report on real-time object detection design. arXiv. https:\/\/doi.org\/10.48550\/arXiv.2211.15444","DOI":"10.48550\/arXiv.2211.15444"},{"key":"17435_CR50","unstructured":"Jocher G, Chaurasia A, Qiu J (2023) YOLO by Ultralytics (Version 8.0.0). https:\/\/github.com\/ultralytics\/ultralytics"},{"key":"17435_CR51","unstructured":"DECI AI (2023) YOLO-NAS. https:\/\/github.com\/Deci-AI\/super-gradients\/blob\/master\/YOLONAS.md"},{"key":"17435_CR52","doi-asserted-by":"publisher","first-page":"271","DOI":"10.1016\/j.biosystemseng.2021.08.015","volume":"210","author":"D Wang","year":"2021","unstructured":"Wang D, He D (2021) Channel pruned YOLO V5s-based deep learning approach for rapid and accurate apple fruitlet detection before fruit thinning. Biosyst Eng 210:271\u2013281. https:\/\/doi.org\/10.1016\/j.biosystemseng.2021.08.015","journal-title":"Biosyst Eng"},{"key":"17435_CR53","doi-asserted-by":"publisher","unstructured":"Wu D, Lv S, Jiang M, Song H (2020) Using channel pruning-based YOLO v4 deep learning algorithm for the real-time and accurate detection of apple flowers in natural environments. Comput Electron Agric 178. https:\/\/doi.org\/10.1016\/J.COMPAG.2020.105742","DOI":"10.1016\/J.COMPAG.2020.105742"},{"key":"17435_CR54","doi-asserted-by":"publisher","first-page":"111808","DOI":"10.1016\/j.postharvbio.2021.111808","volume":"185","author":"Z Wang","year":"2022","unstructured":"Wang Z, Jin L, Wang S, Xu H (2022) Apple stem\/calyx real-time recognition using YOLO-v5 algorithm for fruit automatic loading system. Postharvest Biol Technol 185:111808. https:\/\/doi.org\/10.1016\/j.postharvbio.2021.111808","journal-title":"Postharvest Biol Technol"},{"key":"17435_CR55","doi-asserted-by":"publisher","unstructured":"Yin Y, Li H, Fu W (2020) Faster-YOLO: an accurate and faster object detection method. Digit Sign Process: Rev J 102. https:\/\/doi.org\/10.1016\/J.DSP.2020.102756","DOI":"10.1016\/J.DSP.2020.102756"},{"key":"17435_CR56","unstructured":"Thuan D (2021) Evolution of Yolo algorithm and Yolov5: the state-of-the-art object detection algorithm. Oulu University of Applied Sciences. https:\/\/www.theseus.fi\/handle\/10024\/452552"},{"key":"17435_CR57","doi-asserted-by":"publisher","first-page":"126","DOI":"10.1007\/s44196-023-00302-w","volume":"16","author":"U Sirisha","year":"2023","unstructured":"Sirisha U, Praveen SP, Srinivasu PN et al (2023) Statistical analysis of design aspects of various YOLO-based deep learning models for object detection. Int J Comput Intell Syst 16:126. https:\/\/doi.org\/10.1007\/s44196-023-00302-w","journal-title":"Int J Comput Intell Syst"},{"key":"17435_CR58","doi-asserted-by":"publisher","first-page":"9961","DOI":"10.1007\/s00521-021-05764-7","volume":"33","author":"I Martinez-Alpiste","year":"2021","unstructured":"Martinez-Alpiste I, Golcarenarenji G, Wang Q, Alcaraz-Calero JM (2021) A dynamic discarding technique to increase speed and preserve accuracy for YOLOv3. Neural Comput Appl 33:9961\u20139973. https:\/\/doi.org\/10.1007\/s00521-021-05764-7","journal-title":"Neural Comput Appl"},{"key":"17435_CR59","unstructured":"Tzutalin (2015) LabelImg. Git code. https:\/\/github.com\/HumanSignal\/labelImg"},{"key":"17435_CR60","doi-asserted-by":"publisher","first-page":"103514","DOI":"10.1016\/J.DSP.2022.103514","volume":"126","author":"SSA Zaidi","year":"2022","unstructured":"Zaidi SSA, Ansari MS, Aslam A et al (2022) A survey of modern deep learning based object detection models. Digit Signal Process 126:103514. https:\/\/doi.org\/10.1016\/J.DSP.2022.103514","journal-title":"Digit Signal Process"},{"key":"17435_CR61","doi-asserted-by":"publisher","unstructured":"Rezatofighi H, Tsoi N, Gwak J et al (2019) Generalized intersection over union: a metric and a loss for bounding box regression. In: 2019 IEEE\/CVF conference on computer vision and pattern recognition (CVPR). IEEE, pp 658\u2013666. https:\/\/doi.org\/10.1109\/CVPR.2019.00075","DOI":"10.1109\/CVPR.2019.00075"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-17435-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-17435-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-17435-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,15]],"date-time":"2024-05-15T07:12:39Z","timestamp":1715757159000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-17435-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,4]]},"references-count":61,"journal-issue":{"issue":"17","published-online":{"date-parts":[[2024,5]]}},"alternative-id":["17435"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-17435-x","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,11,4]]},"assertion":[{"value":"3 August 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 September 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 October 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 November 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}