{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T16:47:02Z","timestamp":1776098822139,"version":"3.50.1"},"reference-count":67,"publisher":"MDPI AG","issue":"24","license":[{"start":{"date-parts":[[2021,12,20]],"date-time":"2021-12-20T00:00:00Z","timestamp":1639958400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Current methods of broadcast herbicide application cause a negative environmental and economic impact. Computer vision methods, specifically those related to object detection, have been reported to aid in site-specific weed management procedures for targeted herbicide application within a field. However, a major challenge to developing a weed detection system is the requirement for a properly annotated database to differentiate between weeds and crops under field conditions. This research involved creating an annotated database of 374 red, green, and blue (RGB) color images organized into monocot and dicot weed classes. The images were acquired from corn and soybean research plots located in north-central Indiana using an unmanned aerial system (UAS) flown at 30 and 10 m heights above ground level (AGL). A total of 25,560 individual weed instances were manually annotated. The annotated database consisted of four different subsets (Training Image Sets 1\u20134) to train the You Only Look Once version 3 (YOLOv3) deep learning model for five separate experiments. The best results were observed with Training Image Set 4, consisting of images acquired at 10 m AGL. For monocot and dicot weeds, respectively, an average precision (AP) score of 91.48 % and 86.13% was observed at a 25% IoU threshold (AP @ T = 0.25), as well as 63.37% and 45.13% at a 50% IoU threshold (AP @ T = 0.5). This research has demonstrated a need to develop large, annotated weed databases to evaluate deep learning models for weed identification under field conditions. It also affirms the findings of other limited research studies utilizing object detection for weed identification under field conditions.<\/jats:p>","DOI":"10.3390\/rs13245182","type":"journal-article","created":{"date-parts":[[2021,12,21]],"date-time":"2021-12-21T04:23:47Z","timestamp":1640060627000},"page":"5182","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":53,"title":["Deep Learning-Based Object Detection System for Identifying Weeds Using UAS Imagery"],"prefix":"10.3390","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2216-7875","authenticated-orcid":false,"given":"Aaron","family":"Etienne","sequence":"first","affiliation":[{"name":"Department of Agricultural and Biological Engineering, Purdue University, West Lafayette, IN 47907, USA"}]},{"given":"Aanis","family":"Ahmad","sequence":"additional","affiliation":[{"name":"Elmore Family School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN 47907, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0442-5474","authenticated-orcid":false,"given":"Varun","family":"Aggarwal","sequence":"additional","affiliation":[{"name":"Elmore Family School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN 47907, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8882-0510","authenticated-orcid":false,"given":"Dharmendra","family":"Saraswat","sequence":"additional","affiliation":[{"name":"Department of Agricultural and Biological Engineering, Purdue University, West Lafayette, IN 47907, USA"}]}],"member":"1968","published-online":{"date-parts":[[2021,12,20]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"39","DOI":"10.1016\/j.biosystemseng.2018.03.006","article-title":"Recognising weeds in a maize crop using a random forest machine-learning algorithm and near-infrared snapshot mosaic hyperspectral imagery","volume":"170","author":"Gao","year":"2018","journal-title":"Biosyst. Eng."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"342","DOI":"10.1017\/wet.2017.116","article-title":"Potential corn yield losses from weeds in North America","volume":"32","author":"Soltani","year":"2016","journal-title":"Weed Technol."},{"key":"ref_3","unstructured":"Lingenfelter, D.D., and Hartwig, N.L. (2021, October 06). Introduction to Weeds and Herbicides. Available online: https:\/\/extension.psu.edu\/introduction-to-weeds-and-herbicides."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"533","DOI":"10.1016\/j.asoc.2015.08.027","article-title":"A Semi-supervised system for weed mapping in sunflower crops using unmanned aerial vehicles and a crop row detection method","volume":"37","year":"2015","journal-title":"Appl. Soft Comput. J."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1186\/s12911-018-0631-9","article-title":"Towards improving diagnosis of skin diseases by combining deep neural network and human knowledge","volume":"18","author":"Zhang","year":"2018","journal-title":"BMC Med. Inform. Decis. Mak."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"36501","DOI":"10.1117\/1.JMI.5.3.036501","article-title":"DeepLesion: Automated mining of large-scale lesion annotations and universal lesion detection with deep learning","volume":"5","author":"Yan","year":"2018","journal-title":"J. Med. Imaging"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"721","DOI":"10.1109\/TITB.2009.2017529","article-title":"Overview of advanced computer vision systems for skin lesions characterization","volume":"13","author":"Maglogiannis","year":"2009","journal-title":"IEEE Trans. Inf. Technol. Biomed."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"322","DOI":"10.1007\/978-3-030-44289-7_30","article-title":"Deep learning in breast cancer detection and classification","volume":"1153","author":"Hamed","year":"2020","journal-title":"Adv. Intell. Syst. Comput."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"629","DOI":"10.1007\/978-981-13-3648-5_76","article-title":"A study on computer vision techniques for self-driving cars","volume":"542","author":"Agarwal","year":"2018","journal-title":"Lect. Notes Electr. Eng."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Khan, M., Hassan, S., Ahmed, S.I., and Iqbal, J. (2017, January 9). Stereovision-based real-time obstacle detection scheme for unmanned ground vehicle with steering wheel drive mechanism. Proceedings of the 2017 International Conference on Communication, Computing and Digital Systems, C-CODE, Islamabad, Pakistan.","DOI":"10.1109\/C-CODE.2017.7918961"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"447","DOI":"10.1016\/j.eswa.2017.09.033","article-title":"Survey of computer vision algorithms and applications for unmanned aerial vehicles","volume":"92","year":"2018","journal-title":"Expert Syst. Appl."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"131","DOI":"10.1177\/0025802419893168","article-title":"Facial-recognition algorithms: A literature review","volume":"60","author":"Kaur","year":"2020","journal-title":"Med. Sci. Law"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Zhao, C., and Chen, B. (2019, January 25). Real-time pedestrian detection based on improved YOLO model. Proceedings of the 2019 11th International Conference on Intelligent Human-Machine Systems and Cybernetics, IHMSC, Hangzhou, China.","DOI":"10.1109\/IHMSC.2019.10101"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Heng, L., Meier, L., Tanskanen, P., Fraundorfer, F., and Pollefeys, M. (2011, January 12). Autonomous obstacle avoidance and maneuvering on a vision-guided mav using on-board processing. Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China.","DOI":"10.1109\/ICRA.2011.5980095"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"106081","DOI":"10.1016\/j.compag.2021.106081","article-title":"Performance of deep learning models for classifying and detecting common weeds in corn and soybean production systems","volume":"184","author":"Ahmad","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"1419","DOI":"10.3389\/fpls.2016.01419","article-title":"Using deep learning for image-based plant disease detection","volume":"7","author":"Mohanty","year":"2016","journal-title":"Front. Plant Sci."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"45301","DOI":"10.1109\/ACCESS.2019.2909522","article-title":"PestNet: An end-to-end deep learning approach for large-scale multi-class pest detection and classification","volume":"7","author":"Liu","year":"2019","journal-title":"IEEE Access"},{"key":"ref_18","first-page":"971","article-title":"Plant nutrient deficiency detection using deep convolutional neural network","volume":"13","author":"Wulandhari","year":"2019","journal-title":"ICIC Express Lett."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Luo, L., Liu, W., Lu, Q., Wang, J., Wen, W., Yan, D., Tang, Y., and Gasteratos, A. (2021). Grape berry detection and size measurement based on edge image processing and geometric morphology. Machines, 9.","DOI":"10.3390\/machines9100233"},{"key":"ref_20","unstructured":"Shrestha, D., Steward, B., Kaspar, T., and Robert, P. (2002, January 15). Others determination of early stage corn plant height using stereo vision. Proceedings of the International Conference on Precision Agriculture Abstracts and Proceedings, Minneapolis, MN, USA."},{"key":"ref_21","first-page":"1","article-title":"Computer vision technology in agricultural automation\u2014A review","volume":"7","author":"Tian","year":"2020","journal-title":"Inf. Process. Agric."},{"key":"ref_22","first-page":"243","article-title":"Fruits and vegetables quality evaluation using computer vision: A review","volume":"33","author":"Bhargava","year":"2021","journal-title":"J. King Saud Univ. Comput. Inf. Sci."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Mohidem, N.A., Norasma, N., Ya, C., Shukor Juraimi, A., Fazilah, W., Ilahi, F., Huzaifah, M., Roslim, M., Sulaiman, N., and Saberioon, M. (2021). How can unmanned aerial vehicles be used for detecting weeds in agricultural fields?. Agriculture, 11.","DOI":"10.3390\/agriculture11101004"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"181","DOI":"10.5194\/isprs-archives-XLII-1-181-2018","article-title":"An efficient weed detection procedure using low-cost uav imagery system for precision agriculture applications","volume":"XLII-1","author":"Hassanein","year":"2018","journal-title":"Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"337","DOI":"10.1016\/j.compag.2010.12.011","article-title":"Real-time image processing for crop\/weed discrimination in maize fields","volume":"75","author":"Ribeiro","year":"2011","journal-title":"Comput. Electron. Agric."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"450","DOI":"10.1016\/j.biosystemseng.2011.09.011","article-title":"Weed identification using an Automated Active Shape Matching (AASM) technique","volume":"110","author":"Swain","year":"2011","journal-title":"Biosyst. Eng."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Dadashzadeh, M., Abbaspour-Gilandeh, Y., Mesri-Gundoshmian, T., Sabzi, S., Hern\u00e1ndez-Hern\u00e1ndez, J.L., Hern\u00e1ndez-Hern\u00e1ndez, M., and Ignacio Arribas, J. (2020). Weed classification for site-specific weed management using an automated stereo computer-vision machine-learning system in rice fields. Plants, 9.","DOI":"10.3390\/plants9050559"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"153","DOI":"10.1016\/j.compag.2017.12.032","article-title":"Evaluation of support vector machine and artificial neural networks in weed detection using shape features","volume":"145","author":"Bakhshipour","year":"2018","journal-title":"Comput. Electron. Agric."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"226","DOI":"10.1016\/j.compag.2019.02.005","article-title":"A review on weed detection using ground-based machine vision and image processing techniques","volume":"158","author":"Wang","year":"2019","journal-title":"Comput. Electron. Agric."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"471","DOI":"10.3390\/agriengineering2030032","article-title":"A deep learning approach for weed detection in lettuce crops using multispectral images","volume":"2","author":"Osorio","year":"2020","journal-title":"AgriEngineering"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"2058","DOI":"10.1038\/s41598-018-38343-3","article-title":"DeepWeeds: A multiclass weed species image dataset for deep learning","volume":"9","author":"Olsen","year":"2019","journal-title":"Sci. Rep."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Sa, I., Popovi\u0107, M., Khanna, R., Chen, Z., Lottes, P., Liebisch, F., Nieto, J., Stachniss, C., Walter, A., and Siegwart, R. (2018). WeedMap: A large-scale semantic weed mapping framework using aerial multispectral imaging and deep neural network for precision farming. Remote Sens., 10.","DOI":"10.3390\/rs10091423"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Perez-Ortiz, M., Gutierrez, P.A., Pena, J.M., Torres-Sanchez, J., Lopez-Granados, F., and Hervas-Martinez, C. (2016, January 6\u20139). Machine learning paradigms for weed mapping via unmanned aerial vehicles. Proceedings of the 2016 IEEE Symposium Series on Computational Intelligence, Athens, Greece.","DOI":"10.1109\/SSCI.2016.7849987"},{"key":"ref_34","first-page":"23","article-title":"Weed location and recognition based on UAV imaging and deep learning","volume":"3","author":"Zhang","year":"2020","journal-title":"Int. J. Precis. Agric. Aviat."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"41","DOI":"10.5194\/isprs-annals-IV-2-W3-41-2017","article-title":"Real-time blob-wise sugar beets vs weeds classification for monitoring fields using convolutional neural networks","volume":"4","author":"Milioto","year":"2017","journal-title":"ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Manfreda, S., McCabe, M.F., Miller, P.E., Lucas, R., Madrigal, V.P., Mallinis, G., Dor, E.B., Helman, D., Estes, L., and Ciraolo, G. (2018). On the use of unmanned aerial systems for environmental monitoring. Remote Sens., 10.","DOI":"10.20944\/preprints201803.0097.v1"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"314","DOI":"10.1016\/j.compag.2017.10.027","article-title":"Weed detection in soybean crops using convnets","volume":"143","author":"Pistori","year":"2017","journal-title":"Comput. Electron. Agric."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"e0196302","DOI":"10.1371\/journal.pone.0196302","article-title":"A fully convolutional network for weed mapping of Unmanned Aerial Vehicle (UAV) imagery","volume":"13","author":"Huang","year":"2018","journal-title":"PLoS ONE"},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Sivakumar, A.N.V., Li, J., Scott, S., Psota, E., Jhala, A.J., Luck, J.D., and Shi, Y. (2020). Comparison of object detection and patch-based classification deep learning models on mid-to late-season weed detection in UAV imagery. Remote Sens., 12.","DOI":"10.3390\/rs12132136"},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"14680","DOI":"10.3390\/rs71114680","article-title":"Transferring deep convolutional neural networks for the scene classification of high-resolution remote sensing imagery","volume":"7","author":"Hu","year":"2015","journal-title":"Remote Sens."},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: Towards real-time object detection with region proposal networks","volume":"39","author":"Ren","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016, January 11\u201314). SSD: Single shot multibox detector. Proceedings of the Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.biosystemseng.2019.05.002","article-title":"Maize seedling detection under different growth stages and complex field environments based on an improved faster R\u2013CNN","volume":"184","author":"Quan","year":"2019","journal-title":"Biosyst. Eng."},{"key":"ref_44","unstructured":"Espinoza, M., Le, C.Z., Raheja, A., and Bhandari, S. (2016, January 17). Weed identification and removal using machine learning techniques and unmanned ground vehicles. Proceedings of the SPIE 9866, Autonomous Air and Ground Sensing Systems for Agricultural Optimization and Phenotyping, Baltimore, MD, USA."},{"key":"ref_45","unstructured":"Le, V.N.T., Truong, G., and Alameh, K. (2021, January 13). Detecting weeds from crops under complex field environments based on faster RCNN. Proceedings of the ICCE 2020\u20142020 IEEE 8th International Conference on Communications and Electronics, Phu Quoc Island, Vietnam."},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27\u201330). You only look once: Unified, real-time object detection. Proceedings of the Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Li, M., Zhang, Z., Lei, L., Wang, X., and Guo, X. (2020). Agricultural greenhouses detection in high-resolution satellite images based on convolutional neural networks: Comparison of faster R-CNN, YOLO v3 and SSD. Sensors, 20.","DOI":"10.3390\/s20174938"},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"264","DOI":"10.1007\/s11119-019-09666-6","article-title":"Vegetation detection and discrimination within vegetable plasticulture row-middles using a convolutional neural network","volume":"21","author":"Sharpe","year":"2020","journal-title":"Precis. Agric."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"339","DOI":"10.1016\/j.compag.2018.12.048","article-title":"Development and evaluation of a low-cost and smart technology for precision weed management utilizing artificial intelligence","volume":"157","author":"Partel","year":"2019","journal-title":"Comput. Electron. Agric."},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"419","DOI":"10.5194\/isprs-archives-XLII-2-419-2018","article-title":"Semantic segmentation and unregistered building detection from uav images using a deconvolutional network","volume":"XLII-2","author":"Ham","year":"2018","journal-title":"Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Etienne, A., and Saraswat, D. (2019, January 14). Machine learning approaches to automate weed detection by UAV based sensors. Proceedings of the SPIE 11008, Autonomous Air and Ground Sensing Systems for Agricultural Optimization and Phenotyping IV, Baltimore, MD, USA.","DOI":"10.1117\/12.2520536"},{"key":"ref_52","unstructured":"(2021, October 17). Novus Ag Growth Stages of Corn & Soybeans. Available online: https:\/\/www.novusag.com\/2019\/06\/growth-stages-of-corn-soybeans\/."},{"key":"ref_53","unstructured":"(2021, August 23). Pioneer Staging Corn Growth. Available online: https:\/\/www.pioneer.com\/us\/agronomy\/staging_corn_growth.html."},{"key":"ref_54","unstructured":"Redmon, J., and Farhadi, A. (2018). YOLOv3: An incremental improvement. arXiv."},{"key":"ref_55","unstructured":"Kathuria, A. (2021, November 29). What\u2019s new in YOLO v3? A Review of the YOLO v3 Object Detection. Available online: https:\/\/towardsdatascience.com\/yolo-v3-object-detection-53fb7d3bfe6b."},{"key":"ref_56","unstructured":"Lin, T. (2021, November 29). LabelImg: Graphical Image Annotation Tool. Available online: https:\/\/github.com\/tzutalin\/labelImg."},{"key":"ref_57","unstructured":"Margaritoff, M. (2021, August 08). Drones in Agriculture: How UAVs Make Farming More Efficient. Available online: https:\/\/www.thedrive.com\/tech\/18456\/drones-in-agriculture-how-uavs-make-farming-more-efficient."},{"key":"ref_58","unstructured":"Thomison, P. (2021, August 12). How to Identify Emergence Issues in Corn. Available online: https:\/\/cfaes.osu.edu\/news\/articles\/how-identify-emergence-issues-in-corn."},{"key":"ref_59","unstructured":"Bochkovskiy, A. (2019, May 15). How to Train (to Detect Your Custom Objects). Available online: https:\/\/github.com\/AlexeyAB\/darknet#how-to-train-to-detect-your-custom-objects."},{"key":"ref_60","unstructured":"(2019, March 10). Information Technology at Purdue University Overview of Gilbreth Community Cluster. Available online: https:\/\/www.rcac.purdue.edu\/compute\/gilbreth."},{"key":"ref_61","unstructured":"(2021, February 28). National Pesticide Information Center Dicamba General Fact Sheet. Available online: http:\/\/npic.orst.edu\/factsheets\/dicamba_gen.pdf."},{"key":"ref_62","unstructured":"Aidouni, M. (2021, November 12). Evaluating Object Detection Models: Guide to Performance Metrics. Available online: https:\/\/manalelaidouni.github.io\/Evaluating-Object-Detection-Models-Guide-to-Performance-Metrics.html."},{"key":"ref_63","unstructured":"Hui, J. (2021, October 12). Real-Time Object Detection with YOLO, YOLOv2 and Now YOLOv3. Available online: https:\/\/jonathan-hui.medium.com\/real-time-object-detection-with-yolo-yolov2-28b1b93e2088."},{"key":"ref_64","unstructured":"Bochkovskiy, A., Wang, C.-Y., and Liao, H.-Y.M. (2020). YOLOv4: Optimal speed and accuracy of object detection. arXiv."},{"key":"ref_65","unstructured":"Bochkovskiy, A. (2018, November 18). Yolo Mark Labelling Tool. Available online: https:\/\/github.com\/AlexeyAB\/Yolo_mark."},{"key":"ref_66","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Mark Liao, H.Y., Wu, Y.H., Chen, P.Y., Hsieh, J.W., and Yeh, I.H. (2020, January 16\u201318). CSPNet: A New backbone that can enhance learning capability of CNN. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA.","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"ref_67","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1186\/s40538-021-00217-8","article-title":"Drone and sensor technology for sustainable weed management: A review","volume":"8","author":"Esposito","year":"2021","journal-title":"Chem. Biol. Technol. Agric."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/13\/24\/5182\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:50:00Z","timestamp":1760169000000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/13\/24\/5182"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12,20]]},"references-count":67,"journal-issue":{"issue":"24","published-online":{"date-parts":[[2021,12]]}},"alternative-id":["rs13245182"],"URL":"https:\/\/doi.org\/10.3390\/rs13245182","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,12,20]]}}}