{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T09:48:51Z","timestamp":1762508931413,"version":"build-2065373602"},"reference-count":34,"publisher":"MDPI AG","issue":"7","license":[{"start":{"date-parts":[[2023,3,24]],"date-time":"2023-03-24T00:00:00Z","timestamp":1679616000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62266010","[2019]57","[2019]31"],"award-info":[{"award-number":["62266010","[2019]57","[2019]31"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Cultivation Project of Guizhou University","award":["62266010","[2019]57","[2019]31"],"award-info":[{"award-number":["62266010","[2019]57","[2019]31"]}]},{"name":"Research Project of Guizhou University for Talent Introduction","award":["62266010","[2019]57","[2019]31"],"award-info":[{"award-number":["62266010","[2019]57","[2019]31"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Chili recognition is one of the critical technologies for robots to pick chilies. The robots need locate the fruit. Furthermore, chilies are always planted intensively and their fruits are always clustered. It is a challenge to recognize and locate the chilies that are blocked by branches and leaves, or other chilies. However, little is known about the recognition algorithms considering this situation. Failure to solve this problem will mean that the robot cannot accurately locate and collect chilies, which may even damage the picking robot\u2019s mechanical arm and end effector. Additionally, most of the existing ground target recognition algorithms are relatively complex, and there are many problems, such as numerous parameters and calculations. Many of the existing models have high requirements for hardware and poor portability. It is very difficult to perform these algorithms if the picking robots have limited computing and battery power. In view of these practical issues, we propose a target recognition-location scheme GNPD-YOLOv5s based on improved YOLOv5s in order to automatically identify the occluded and non-occluded chilies. Firstly, the lightweight optimization for Ghost module is introduced into our scheme. Secondly, pruning and distilling the model is designed to further reduce the number of parameters. Finally, the experimental data show that compared with the YOLOv5s model, the floating point operation number of the GNPD-YOLOv5s scheme is reduced by 40.9%, the model size is reduced by 46.6%, and the reasoning speed is accelerated from 29 ms\/frame to 14 ms\/frame. At the same time, the Mean Accuracy Precision (MAP) is reduced by 1.3%. Our model implements a lightweight network model and target recognition in the dense environment at a small cost. In our locating experiments, the maximum depth locating chili error is 1.84 mm, which meets the needs of a chili picking robot for chili recognition.<\/jats:p>","DOI":"10.3390\/s23073408","type":"journal-article","created":{"date-parts":[[2023,3,24]],"date-time":"2023-03-24T03:16:46Z","timestamp":1679627806000},"page":"3408","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["Real-Time Recognition and Localization Based on Improved YOLOv5s for Robot\u2019s Picking Clustered Fruits of Chilies"],"prefix":"10.3390","volume":"23","author":[{"given":"Song","family":"Zhang","sequence":"first","affiliation":[{"name":"College of Big Data and Information Engineering, Guizhou University, Guiyang 550025, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6387-785X","authenticated-orcid":false,"given":"Mingshan","family":"Xie","sequence":"additional","affiliation":[{"name":"College of Big Data and Information Engineering, Guizhou University, Guiyang 550025, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,3,24]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"105687","DOI":"10.1016\/j.compag.2020.105687","article-title":"Application of consumer RGB-D cameras for fruit recognition and localization in field: A critical review. Comput","volume":"177","author":"Fu","year":"2020","journal-title":"Electron. Agric."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"31","DOI":"10.1007\/s00138-020-01081-6","article-title":"Recognition of tomato organs based on convolutional neural network under the overlap and occlusion backgrounds","volume":"31","author":"Sun","year":"2020","journal-title":"Mach. Vis. Appl."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"113223","DOI":"10.1016\/j.indcrop.2020.113223","article-title":"Distinguishing seedling volunteer corn from soybean through greenhouse color, color-infrared, and fused images using machine and deep learning","volume":"161","author":"Flores","year":"2021","journal-title":"Ind. Crops Prod."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"26","DOI":"10.1016\/j.compag.2015.10.022","article-title":"Apple crop-load estimation with over-the-row machine vision system","volume":"120","author":"Gongal","year":"2016","journal-title":"Comput. Electron. Agric."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"111201","DOI":"10.1016\/j.postharvbio.2020.111201","article-title":"Evaluation of image processing technique as an expert system in mulberry fruit grading based on ripeness level using artificial neural networks (ANNs) and support vector machine (SVM)","volume":"166","author":"Azarmdel","year":"2020","journal-title":"Postharvest Biol. Technol."},{"key":"ref_6","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016). Proceedings of the IEEE Conference on Computer Vision and Pattern 2016 Recognition (CVPR), Las Vegas, NV, USA, 26 June\u20131 July 2016, IEEE."},{"key":"ref_7","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., and Berg, A.C. (2016). Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11\u201314 October 2016, IEEE."},{"key":"ref_8","first-page":"2999","article-title":"Focal Loss for Dense Object Recognition","volume":"99","author":"Lin","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"207","DOI":"10.1016\/j.trc.2019.09.012","article-title":"Policy processes and recommendations for Unmanned Aerial System operations near roadways based on visual attention of drivers","volume":"108","author":"Barlow","year":"2019","journal-title":"Transp. Res. Part C Emerg. Technol."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Chen, T., Wang, Z., Li, G., and Lin, L. (2018, January 2\u20137). Recurrent attentional reinforcement learning for multi-label image recognition. Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA.","DOI":"10.1609\/aaai.v32i1.12281"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Zhu, C., Wu, C., Li, Y., Hu, S., and Gong, H. (2022). Spatial Location of Sugarcane Node for Binocular Vision-Based Harvesting Robots Based on Improved YOLOv4. Appl. Sci., 12.","DOI":"10.3390\/app12063088"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Fuentes, A., Yoon, S., Kim, S.C., and Park, D.S. (2017). A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition. Sensors, 17.","DOI":"10.3390\/s17092022"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"139635","DOI":"10.1109\/ACCESS.2019.2942144","article-title":"Cucumber fruits recognition in greenhouses based on instance segmentation","volume":"7","author":"Liu","year":"2019","journal-title":"IEEE Access"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"5916818","DOI":"10.1155\/2020\/5916818","article-title":"Research and verification of convolutional neural network lightweight in BCI","volume":"2020","author":"Xu","year":"2020","journal-title":"Comput. Math. Methods Med."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"106503","DOI":"10.1016\/j.compag.2021.106503","article-title":"Fast and accurate green pepper detection in complex backgrounds via an improved Yolov4-tiny model","volume":"191","author":"Li","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"ref_16","unstructured":"Liu, Z., Sun, M., Zhou, T., Huang, G., and Darrell, T. (2018). Rethinking the value of network pruning. arXiv."},{"key":"ref_17","unstructured":"Yin, P., Lyu, J., Zhang, S., Osher, S., Qi, Y., and Xin, J. (2019). Understanding straight-through estimator in training activation quantized neural nets. arXiv."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"1789","DOI":"10.1007\/s11263-021-01453-z","article-title":"Knowledge distillation: A survey","volume":"129","author":"Gou","year":"2021","journal-title":"Int. J. Comput. Vis."},{"key":"ref_19","first-page":"506","article-title":"Fixed point implementation of tiny-yolo-v2 using opencl on fpga","volume":"9","author":"Yap","year":"2018","journal-title":"Int. J. Adv. Comput. Sci. Appl."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"17","DOI":"10.1016\/j.ijleo.2019.02.038","article-title":"An improved tiny-yolov3 pedestrian recognition algorithm","volume":"183","author":"Yi","year":"2019","journal-title":"Optik"},{"key":"ref_21","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. (2017). Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and Chen, L.-C. (2018, January 18\u201323). Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref_23","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., and Vasudevan, V. (November, January 27). Searching for mobilenetv3. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Korea."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., and Le, Q.V. (2020, January 13\u201319). Efficientdet: Scalable and efficient object recognition. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., and Xu, C. (2020, January 13\u201319). Ghostnet: More features from cheap operations. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., and Sun, J. (2018, January 18\u201323). Shufflenet: An extremely efficient convolutional neural network for mobile devices. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00716"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.T., and Sun, J. (2018, January 8\u201314). Shufflenet v2: Practical guidelines for efficient cnn architecture design. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"ref_28","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., and Keutzer, K. (2016). SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"585","DOI":"10.1016\/j.compag.2018.12.006","article-title":"Current and future applications of statistical machine learning algorithms for agricultural machine vision systems","volume":"156","author":"Rehman","year":"2019","journal-title":"Comput. Electron. Agric."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"2292","DOI":"10.1364\/OE.27.002292","article-title":"A high precision 3D reconstruction method for bend tube axis based on binocular stereo vision","volume":"27","author":"Sun","year":"2019","journal-title":"Opt. Express"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"111615","DOI":"10.1016\/j.postharvbio.2021.111615","article-title":"A robot system for the autorecognition and classification of apple internal quality attributes","volume":"180","author":"Zhao","year":"2021","journal-title":"Postharvest Biol. Technol."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"107036","DOI":"10.1016\/j.comnet.2019.107036","article-title":"Faster R-CNN for multi-class fruit recognition using a robotic vision system","volume":"168","author":"Wan","year":"2020","journal-title":"Comput. Netw."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"1729881420925310","DOI":"10.1177\/1729881420925310","article-title":"Apple harvesting robot under information technology: A review","volume":"17","author":"Jia","year":"2020","journal-title":"Int. J. Adv. Robot. Syst."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"130","DOI":"10.1016\/j.compind.2018.03.017","article-title":"A vision methodology for harvesting robot to detect cutting points on peduncles of double overlapping grape clusters in a vineyard","volume":"99","author":"Luo","year":"2018","journal-title":"Comput. Ind."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/23\/7\/3408\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T19:01:57Z","timestamp":1760122917000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/23\/7\/3408"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,24]]},"references-count":34,"journal-issue":{"issue":"7","published-online":{"date-parts":[[2023,4]]}},"alternative-id":["s23073408"],"URL":"https:\/\/doi.org\/10.3390\/s23073408","relation":{},"ISSN":["1424-8220"],"issn-type":[{"type":"electronic","value":"1424-8220"}],"subject":[],"published":{"date-parts":[[2023,3,24]]}}}