{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T17:34:52Z","timestamp":1763141692981,"version":"build-2065373602"},"reference-count":51,"publisher":"MDPI AG","issue":"23","license":[{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100006206","name":"Czech University of Life Sciences Prague","doi-asserted-by":"publisher","award":["2019B0009 \u2013 Life Sciences 4.0","CZ.02.2.69\/0.0\/0.0\/16_027\/0008366"],"award-info":[{"award-number":["2019B0009 \u2013 Life Sciences 4.0","CZ.02.2.69\/0.0\/0.0\/16_027\/0008366"]}],"id":[{"id":"10.13039\/501100006206","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Government of India","award":["\u25cf\tEarly Career Research Award from Department of Science and Technology"],"award-info":[{"award-number":["\u25cf\tEarly Career Research Award from Department of Science and Technology"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>This study tested whether machine learning (ML) methods can effectively separate individual plants from complex 3D canopy laser scans as a prerequisite to analyzing particular plant features. For this, we scanned mung bean and chickpea crops with PlantEye (R) laser scanners. Firstly, we segmented the crop canopies from the background in 3D space using the Region Growing Segmentation algorithm. Then, Convolutional Neural Network (CNN) based ML algorithms were fine-tuned for plant counting. Application of the CNN-based (Convolutional Neural Network) processing architecture was possible only after we reduced the dimensionality of the data to 2D. This allowed for the identification of individual plants and their counting with an accuracy of 93.18% and 92.87% for mung bean and chickpea plants, respectively. These steps were connected to the phenotyping pipeline, which can now replace manual counting operations that are inefficient, costly, and error-prone. The use of CNN in this study was innovatively solved with dimensionality reduction, addition of height information as color, and consequent application of a 2D CNN-based approach. We found there to be a wide gap in the use of ML on 3D information. This gap will have to be addressed, especially for more complex plant feature extractions, which we intend to implement through further research.<\/jats:p>","DOI":"10.3390\/s21238022","type":"journal-article","created":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T03:12:40Z","timestamp":1638328360000},"page":"8022","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Machine Learning-Based Plant Detection Algorithms to Automate Counting Tasks Using 3D Canopy Scans"],"prefix":"10.3390","volume":"21","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9801-8986","authenticated-orcid":false,"given":"Serkan","family":"Kartal","sequence":"first","affiliation":[{"name":"Department of Computer Engineering, Faculty of Engineering, Cukurova University, Adana 01330, Turkey"}]},{"given":"Sunita","family":"Choudhary","sequence":"additional","affiliation":[{"name":"System Analysis for Climate Smart Agriculture (SACSA), ISD, International Crops Research Institute for the Semi-Arid Tropics, Patancheru 5023204, Telangana, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4593-2306","authenticated-orcid":false,"given":"Jan","family":"Masner","sequence":"additional","affiliation":[{"name":"Department of Information Technologies, Faculty of Economics and Management, Czech University of Life Sciences Prague, Kam\u00fdck\u00e1 129, 165 00 Prague, Czech Republic"}]},{"given":"Jana","family":"Kholov\u00e1","sequence":"additional","affiliation":[{"name":"System Analysis for Climate Smart Agriculture (SACSA), ISD, International Crops Research Institute for the Semi-Arid Tropics, Patancheru 5023204, Telangana, India"}]},{"given":"Michal","family":"Sto\u010des","sequence":"additional","affiliation":[{"name":"Department of Information Technologies, Faculty of Economics and Management, Czech University of Life Sciences Prague, Kam\u00fdck\u00e1 129, 165 00 Prague, Czech Republic"}]},{"given":"Priyanka","family":"Gattu","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering, Indian Institute of Technology Hyderabad, Sangareddy 502285, Telangana, India"}]},{"given":"Stefan","family":"Schwartz","sequence":"additional","affiliation":[{"name":"Phenospex B. V., Jan Campertstraat 11, 6416 SG Heerlen, The Netherlands"}]},{"given":"Ewaut","family":"Kissel","sequence":"additional","affiliation":[{"name":"Phenospex B. V., Jan Campertstraat 11, 6416 SG Heerlen, The Netherlands"}]}],"member":"1968","published-online":{"date-parts":[[2021,12,1]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"R770","DOI":"10.1016\/j.cub.2017.05.055","article-title":"Plant phenomics, from sensors to knowledge","volume":"27","author":"Tardieu","year":"2017","journal-title":"Curr. Biol."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"20078","DOI":"10.3390\/s141120078","article-title":"A Review of Imaging Techniques for Plant Phenotyping","volume":"14","author":"Li","year":"2014","journal-title":"Sensors"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"1907","DOI":"10.3389\/fpls.2020.618342","article-title":"Editorial: Phenotyping; From Plant, to Data, to Impact and Highlights of the International Plant Phenotyping Symposium-IPPS 2018","volume":"11","author":"Pommier","year":"2020","journal-title":"Front. Plant Sci."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"5158","DOI":"10.1093\/jxb\/erab226","article-title":"In pursuit of a better world: Crop improvement and the CGIAR","volume":"72","author":"Urban","year":"2021","journal-title":"J. Exp. Bot."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"5581","DOI":"10.1093\/jxb\/erv251","article-title":"LeasyScan: A novel concept combining 3D imaging and lysimetry for high-throughput phenotyping of traits controlling plant water budget","volume":"66","author":"Vadez","year":"2015","journal-title":"J. Exp. Bot."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"635","DOI":"10.1016\/j.tplants.2011.09.005","article-title":"Phenomics\u2014Technologies to relieve the phenotyping bottleneck","volume":"16","author":"Furbank","year":"2011","journal-title":"Trends Plant Sci."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"73","DOI":"10.1016\/j.pbi.2014.02.002","article-title":"TraitCapture: Genomic and environment modelling of plant phenomic data","volume":"18","author":"Brown","year":"2014","journal-title":"Curr. Opin. Plant. Biol."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"143","DOI":"10.1071\/FP16163","article-title":"Field Scanalyzer: An automated robotic field phenotyping platform for detailed crop monitoring","volume":"44","author":"Virlet","year":"2016","journal-title":"Funct. Plant Biol."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"267","DOI":"10.1146\/annurev-arplant-050312-120137","article-title":"Future Scenarios for Plant Phenotyping","volume":"64","author":"Fiorani","year":"2013","journal-title":"Annu. Rev. Plant Biol."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.eja.2012.05.006","article-title":"Designing crops for new challenges","volume":"42","author":"Tardieu","year":"2012","journal-title":"Eur. J. Agron."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"733","DOI":"10.1146\/annurev-arplant-042817-040218","article-title":"The Physiological Basis of Drought Tolerance in Crop Plants: A Scenario-Dependent Probabilistic Approach","volume":"69","author":"Tardieu","year":"2018","journal-title":"Annu. Rev. Plant Biol."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"1019","DOI":"10.1071\/FP13355","article-title":"Modelling the effect of plant water use traits on yield and stay-green expression in sorghum","volume":"41","author":"Murugesan","year":"2014","journal-title":"Funct. Plant Biol."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Sivasakthi, K., Thudi, M., Tharanya, M., Kale, S.M., Kholov\u00e1, J., Halime, M.H., Jaganathan, D., Baddam, R., Thirunalasundari, T., and Gaur, P.M. (2018). Plant vigour QTLs co-map with an earlier reported QTL hotspot for drought tolerance while water saving QTLs map in other regions of the chickpea genome. BMC Plant Biol., 18.","DOI":"10.1186\/s12870-018-1245-1"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Sivasakthi, K., Marques, E., Kalungwana, N., Carrasquilla-Garcia, N., Chang, P.L., Bergmann, E.M., Bueno, E., Cordeiro, M., Sani, S.G.A., and Udupa, S.M. (2019). Functional Dissection of the Chickpea (Cicer arietinum L.) Stay-Green Phenotype Associated with Molecular Variation at an Ortholog of Mendel\u2019s I Gene for Cotyledon Color: Implications for Crop Production and Carotenoid Biofortification. Int. J. Mol. Sci., 20.","DOI":"10.3390\/ijms20225562"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1509","DOI":"10.1007\/s00122-018-3094-6","article-title":"Quantitative trait loci (QTLs) for water use and crop production traits co-locate with major QTL for tolerance to water deficit in a fine-mapping population of pearl millet (Pennisetum glaucum L. R.Br.)","volume":"131","author":"Tharanya","year":"2018","journal-title":"Theor. Appl. Genet."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"552509","DOI":"10.3389\/fpls.2020.552509","article-title":"SpaTemHTP: A Data Analysis Pipeline for Efficient Processing and Utilization of Temporal High-Throughput Phenotyping Data","volume":"11","author":"Kar","year":"2020","journal-title":"Front. Plant Sci."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"140","DOI":"10.1186\/s13007-020-00680-8","article-title":"Automated discretization of \u2018transpiration restriction to increasing VPD\u2019 features from outdoors high-throughput phenotyping data","volume":"16","author":"Kar","year":"2020","journal-title":"Plant Methods"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1186\/1746-4811-10-9","article-title":"Rapid determination of leaf area and plant height by using light curtain arrays in four species with contrasting shoot architecture","volume":"10","author":"Fanourakis","year":"2014","journal-title":"Plant Methods"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"gix083","DOI":"10.1093\/gigascience\/gix083","article-title":"Deep machine learning provides state-of-the-art performance in image-based plant phenotyping","volume":"6","author":"Pound","year":"2017","journal-title":"GigaScience"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"418","DOI":"10.1016\/j.compag.2016.07.003","article-title":"Deep learning for plant identification using vein morphological patterns","volume":"127","author":"Grinblat","year":"2016","journal-title":"Comput. Electron. Agric."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"7361042","DOI":"10.1155\/2017\/7361042","article-title":"Deep Learning for Plant Identification in Natural Environment","volume":"2017","author":"Sun","year":"2017","journal-title":"Comput. Intell. Neurosci."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"11149","DOI":"10.1016\/j.eswa.2012.03.040","article-title":"Support Vector Machines for crop\/weeds identification in maize fields","volume":"39","author":"Guerrero","year":"2012","journal-title":"Expert Syst. Appl."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"521","DOI":"10.1016\/j.patcog.2007.07.007","article-title":"A vision-based method for weeds identification through the Bayesian decision theory","volume":"41","author":"Tellaeche","year":"2008","journal-title":"Pattern Recognit."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"113","DOI":"10.1016\/j.agrformet.2011.10.014","article-title":"An alternative method using digital cameras for continuous monitoring of crop status","volume":"154","author":"Sakamoto","year":"2012","journal-title":"Agric. For. Meteorol."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"19","DOI":"10.1016\/j.biosystemseng.2015.01.008","article-title":"Multi-temporal imaging using an unmanned aerial vehicle for monitoring a sunflower crop","volume":"132","author":"Vega","year":"2015","journal-title":"Biosyst. Eng."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1016\/j.biosystemseng.2013.08.011","article-title":"An automated growth measurement system for leafy vegetables","volume":"117","author":"Yeh","year":"2014","journal-title":"Biosyst. Eng."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"162","DOI":"10.1016\/j.biosystemseng.2013.03.009","article-title":"Citrus yield estimation based on images processed by an Android mobile phone","volume":"115","author":"Gong","year":"2013","journal-title":"Biosyst. Eng."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"57","DOI":"10.1016\/j.compag.2012.11.009","article-title":"Estimation of mango crop yield using image analysis\u2014Segmentation method","volume":"91","author":"Payne","year":"2013","journal-title":"Comput. Electron. Agric."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1016\/j.biosystemseng.2013.05.010","article-title":"Automatic detection of tulip breaking virus (TBV) in tulip fields using machine vision","volume":"117","author":"Polder","year":"2014","journal-title":"Biosyst. Eng."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"13","DOI":"10.1016\/j.biosystemseng.2014.11.013","article-title":"An evaluation of a vision-based sensor performance in Huanglongbing disease identification","volume":"130","author":"Pourreza","year":"2015","journal-title":"Biosyst. Eng."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"94","DOI":"10.1016\/j.biosystemseng.2013.09.003","article-title":"Automatic corn (Zea mays) kernel inspection system using novelty detection based on principal component analysis","volume":"117","author":"Potter","year":"2014","journal-title":"Biosyst. Eng."},{"key":"ref_32","first-page":"51","article-title":"Automated Wildlife Recognition","volume":"10","year":"2018","journal-title":"Agris-Line Pap. Econ. Inform."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"70","DOI":"10.1016\/j.compag.2018.02.016","article-title":"Deep learning in agriculture: A survey","volume":"147","author":"Kamilaris","year":"2018","journal-title":"Comput. Electron. Agric."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"109","DOI":"10.2480\/agrmet.D-18-00012","article-title":"Automatic individual tree detection and canopy segmentation from three-dimensional point cloud images obtained from ground-based lidar","volume":"74","author":"Itakura","year":"2018","journal-title":"J. Agric. Meteorol."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.isprsjprs.2018.12.015","article-title":"Automated detection and measurement of individual sorghum panicles using density-based clustering of terrestrial lidar data","volume":"149","author":"Malambo","year":"2019","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"265","DOI":"10.1016\/j.robot.2011.02.011","article-title":"Plant detection and mapping for agricultural robots using a 3D LIDAR sensor","volume":"59","author":"Weiss","year":"2011","journal-title":"Robot. Auton. Syst."},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"2275","DOI":"10.1109\/TIP.2009.2025555","article-title":"Automatic Image Segmentation by Dynamic Region Growth and Multiresolution Merging","volume":"18","author":"Ugarriza","year":"2009","journal-title":"IEEE Trans. Image Process."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"519","DOI":"10.21608\/mjeer.2017.63627","article-title":"A Survey of RANSAC enhancements for Plane Detection in 3D Point Clouds","volume":"26","author":"Zeineldin","year":"2017","journal-title":"Menoufia J. Electron. Eng. Res."},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"345","DOI":"10.1007\/s13218-010-0059-6","article-title":"Semantic 3D Object Maps for Everyday Manipulation in Human Living Environments","volume":"24","author":"Rusu","year":"2010","journal-title":"KI K\u00fcnstliche Intell."},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks","volume":"39","author":"Ren","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2015, January 7\u201312). Going deeper with convolutions. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_42","unstructured":"Maturana, D., and Scherer, S. (October, January 25). VoxNet: VoxNet: A 3D convolutional neural network for real-time object recognition. Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Hamburg, Germany."},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"22","DOI":"10.1186\/s40537-018-0131-x","article-title":"A hybrid framework combining background subtraction and deep neural networks for rapid person detection","volume":"5","author":"Kim","year":"2018","journal-title":"J. Big Data"},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Mohamed, S.S., Tahir, N.M., and Adnan, R. (2010, January 21\u201323). Background modelling and background subtraction performance for object detection. Proceedings of the 2010 6th International Colloquium on Signal Processing and Its Applications (CSPA 2010), Malacca, Malaysia.","DOI":"10.1109\/CSPA.2010.5545291"},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"3244","DOI":"10.1109\/TVCG.2018.2866793","article-title":"VERAM: Rapid determination of leaf area and plant height by using light curtain arrays in four species with contrasting shoot architecture","volume":"25","author":"Chen","year":"2019","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"ref_46","unstructured":"Qi, C.R., Su, H., Mo, K., and Guibas, L.J. (2017, January 21\u201326). Pointnet: Deep learning on point sets for 3D classification and segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA."},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Yavartanoo, M., Kim, E.Y., and Lee, K.M. (2019). SPNet: Deep 3D object classification and retrieval using stereographic projection. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Springer.","DOI":"10.1007\/978-3-030-20873-8_44"},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"38","DOI":"10.1109\/MGRS.2019.2937630","article-title":"Linking Points with Labels in 3D: A Review of Point Cloud Semantic Segmentation","volume":"8","author":"Xie","year":"2020","journal-title":"IEEE Geosci. Remote Sens. Mag."},{"key":"ref_49","unstructured":"Ioffe, S., and Szegedy, C. (2015, January 7\u20139). Batch normalization: Accelerating deep network training by reducing internal covariate shift. Proceedings of the 32nd International Conference on Machine Learning (ICML), Lille, France."},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioe, S., Shlens, J., and Wojna, Z. (2016, January 27\u201330). Rethinking the inception architecture for computer vision. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.308"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X.Y., Ren, S.Q., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/21\/23\/8022\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:38:20Z","timestamp":1760168300000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/21\/23\/8022"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12,1]]},"references-count":51,"journal-issue":{"issue":"23","published-online":{"date-parts":[[2021,12]]}},"alternative-id":["s21238022"],"URL":"https:\/\/doi.org\/10.3390\/s21238022","relation":{},"ISSN":["1424-8220"],"issn-type":[{"type":"electronic","value":"1424-8220"}],"subject":[],"published":{"date-parts":[[2021,12,1]]}}}