{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T05:34:46Z","timestamp":1772775286962,"version":"3.50.1"},"reference-count":53,"publisher":"MDPI AG","issue":"17","license":[{"start":{"date-parts":[[2020,8,26]],"date-time":"2020-08-26T00:00:00Z","timestamp":1598400000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100005672","name":"Funda\u00e7\u00e3o de Apoio ao Desenvolvimento do Ensino, Ci\u00eancia e Tecnologia do Estado de Mato Grosso do Sul","doi-asserted-by":"publisher","award":["59\/300.075\/2015"],"award-info":[{"award-number":["59\/300.075\/2015"]}],"id":[{"id":"10.13039\/501100005672","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003593","name":"Conselho Nacional de Desenvolvimento Cient\u00edfico e Tecnol\u00f3gico","doi-asserted-by":"publisher","award":["433783\/2018-4; 303559\/2019-5"],"award-info":[{"award-number":["433783\/2018-4; 303559\/2019-5"]}],"id":[{"id":"10.13039\/501100003593","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Monitoring biomass of forages in experimental plots and livestock farms is a time-consuming, expensive, and biased task. Thus, non-destructive, accurate, precise, and quick phenotyping strategies for biomass yield are needed. To promote high-throughput phenotyping in forages, we propose and evaluate the use of deep learning-based methods and UAV (Unmanned Aerial Vehicle)-based RGB images to estimate the value of biomass yield by different genotypes of the forage grass species Panicum maximum Jacq. Experiments were conducted in the Brazilian Cerrado with 110 genotypes with three replications, totaling 330 plots. Two regression models based on Convolutional Neural Networks (CNNs) named AlexNet and ResNet18 were evaluated, and compared to VGGNet\u2014adopted in previous work in the same thematic for other grass species. The predictions returned by the models reached a correlation of 0.88 and a mean absolute error of 12.98% using AlexNet considering pre-training and data augmentation. This proposal may contribute to forage biomass estimation in breeding populations and livestock areas, as well as to reduce the labor in the field.<\/jats:p>","DOI":"10.3390\/s20174802","type":"journal-article","created":{"date-parts":[[2020,8,26]],"date-time":"2020-08-26T09:05:37Z","timestamp":1598432737000},"page":"4802","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":77,"title":["Deep Learning Applied to Phenotyping of Biomass in Forages with UAV-Based RGB Imagery"],"prefix":"10.3390","volume":"20","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2847-2240","authenticated-orcid":false,"given":"Wellington","family":"Castro","sequence":"first","affiliation":[{"name":"Faculty of Computer Science, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9096-6866","authenticated-orcid":false,"given":"Jos\u00e9","family":"Marcato Junior","sequence":"additional","affiliation":[{"name":"Faculty of Engineering, Architecture and Urbanism and Geography, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8212-0828","authenticated-orcid":false,"given":"Caio","family":"Polidoro","sequence":"additional","affiliation":[{"name":"Faculty of Computer Science, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0258-536X","authenticated-orcid":false,"given":"Lucas Prado","family":"Osco","sequence":"additional","affiliation":[{"name":"Faculty of Engineering, Architecture and Urbanism, University of Western S\u00e3o Paulo, Presidente Prudente 19067175, SP, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8815-6653","authenticated-orcid":false,"given":"Wesley","family":"Gon\u00e7alves","sequence":"additional","affiliation":[{"name":"Faculty of Engineering, Architecture and Urbanism and Geography, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4704-066X","authenticated-orcid":false,"given":"Lucas","family":"Rodrigues","sequence":"additional","affiliation":[{"name":"Faculty of Computer Science, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5324-906X","authenticated-orcid":false,"given":"Mateus","family":"Santos","sequence":"additional","affiliation":[{"name":"Embrapa Beef Cattle, Brazilian Agricultural Research Corporation, Campo Grande 79106550, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9436-3678","authenticated-orcid":false,"given":"Liana","family":"Jank","sequence":"additional","affiliation":[{"name":"Embrapa Beef Cattle, Brazilian Agricultural Research Corporation, Campo Grande 79106550, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5490-4959","authenticated-orcid":false,"given":"Sanzio","family":"Barrios","sequence":"additional","affiliation":[{"name":"Embrapa Beef Cattle, Brazilian Agricultural Research Corporation, Campo Grande 79106550, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3054-5127","authenticated-orcid":false,"given":"Cacilda","family":"Valle","sequence":"additional","affiliation":[{"name":"Embrapa Beef Cattle, Brazilian Agricultural Research Corporation, Campo Grande 79106550, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8351-846X","authenticated-orcid":false,"given":"Rosangela","family":"Sime\u00e3o","sequence":"additional","affiliation":[{"name":"Embrapa Beef Cattle, Brazilian Agricultural Research Corporation, Campo Grande 79106550, MS, Brazil"}]},{"given":"Camilo","family":"Carromeu","sequence":"additional","affiliation":[{"name":"Embrapa Beef Cattle, Brazilian Agricultural Research Corporation, Campo Grande 79106550, MS, Brazil"}]},{"given":"Eloise","family":"Silveira","sequence":"additional","affiliation":[{"name":"Faculty of Engineering, Architecture and Urbanism and Geography, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8341-3203","authenticated-orcid":false,"given":"L\u00facio Andr\u00e9 de Castro","family":"Jorge","sequence":"additional","affiliation":[{"name":"Embrapa Instrumentation, S\u00e3o Carlos 13560970, SP, Brazil"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4471-0886","authenticated-orcid":false,"given":"Edson","family":"Matsubara","sequence":"additional","affiliation":[{"name":"Faculty of Computer Science, Federal University of Mato Grosso do Sul, Campo Grande 79070900, MS, Brazil"}]}],"member":"1968","published-online":{"date-parts":[[2020,8,26]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"10395","DOI":"10.3390\/rs61110395","article-title":"Estimating Biomass of Barley Using Crop Surface Models (CSMs) Derived from UAV-Based RGB Imaging","volume":"6","author":"Bendig","year":"2014","journal-title":"Remote Sens."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Gebremedhin, A., Badenhorst, P.E., Wang, J., Spangenberg, G.C., and Smith, K.F. (2019). Prospects for measurement of dry matter yield in forage breeding programs using sensor technologies. Agronomy, 9.","DOI":"10.3390\/agronomy9020065"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"111402","DOI":"10.1016\/j.rse.2019.111402","article-title":"Remote sensing for agricultural applications: A meta-review","volume":"236","author":"Weiss","year":"2020","journal-title":"Remote Sens. Environ."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Osco, L.P., Ramos, A.P.M., Pereira, D.R., Moriya, \u00c9.A.S., Imai, N.N., Matsubara, E.T., Estrabis, N., de Souza, M., Junior, J.M., and Gon\u00e7alves, W.N. (2019). Predicting canopy nitrogen content in citrus-trees using random forest algorithm associated to spectral vegetation indices from UAV-imagery. Remote Sens., 11.","DOI":"10.3390\/rs11242925"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"D\u2019Oliveira, M.V.N., Broadbent, E.N., Oliveira, L.C., Almeida, D.R.A., Papa, D.A., Ferreira, M.E., Zambrano, A.M.A., Silva, C.A., Avino, F.S., and Prata, G.A. (2020). Aboveground Biomass Estimation in Amazonian Tropical Forests: A Comparison of Aircraft- and GatorEye UAV-borne LiDAR Data in the Chico Mendes Extractive Reserve in Acre, Brazil. Remote Sens., 12.","DOI":"10.3390\/rs12111754"},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Miyoshi, G.T., Arruda, M.D.S., Osco, L.P., Junior, J.M., Gon\u00e7alves, D.N., Imai, N.N., Tommaselli, A.M.G., Honkavaara, E., and Gon\u00e7alves, W.N. (2020). A novel deep learning method to identify single tree species in UAV-based hyperspectral images. Remote Sens., 12.","DOI":"10.3390\/rs12081294"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"036003","DOI":"10.1117\/1.JRS.11.036003","article-title":"Evaluating remotely sensed plant count accuracy with differing unmanned aircraft system altitudes, physical canopy separations, and ground covers","volume":"11","author":"Leiva","year":"2017","journal-title":"J. Appl. Remote Sens."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"243","DOI":"10.1080\/15481603.2018.1426091","article-title":"Comparing fully convolutional networks, random forest, support vector machine, and patch-based deep convolutional neural networks for object-based wetland mapping using images from small unmanned aircraft system","volume":"55","author":"Liu","year":"2018","journal-title":"GISci. Remote Sens."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Abdulridha, J., Batuman, O., and Ampatzidis, Y. (2019). UAV-based remote sensing technique to detect citrus canker disease utilizing hyperspectral imaging and machine learning. Remote Sens., 11.","DOI":"10.3390\/rs11111373"},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1016\/j.agsy.2019.03.015","article-title":"Machine learning-based integration of remotely-sensed drought factors can improve the estimation of agricultural drought in South-Eastern Australia","volume":"173","author":"Feng","year":"2019","journal-title":"Agric. Syst."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Osco, L.P., Ramos, A.P.M., Pinheiro, M.M.F., Moriya, \u00c9.A.S., Imai, N.N., Estrabis, N., Ianczyk, F., de Ara\u00fajo, F.F., Liesenberg, V., and de Castro Jorge, L.A. (2020). A machine learning framework to predict nutrient content in valencia-orange leaf hyperspectral measurements. Remote Sens., 12.","DOI":"10.3390\/rs12060906"},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1109\/MGRS.2016.2616418","article-title":"Advanced Spectral Classifiers for Hyperspectral Images: A review","volume":"5","author":"Ghamisi","year":"2017","journal-title":"IEEE Geosci. Remote Sens. Mag."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Al-Saffar, A.A.M., Tao, H., and Talab, M.A. (2017, January 23\u201324). Review of deep convolution neural network in image classification. Proceedings of the 2017 International Conference on Radar, Antenna, Microwave, Electronics, and Telecommunications (ICRAMET), Jakarta, Indonesia.","DOI":"10.1109\/ICRAMET.2017.8253139"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1016\/j.isprsjprs.2017.05.002","article-title":"Simultaneous extraction of roads and buildings in remote sensing imagery with convolutional neural networks","volume":"130","author":"Alshehhi","year":"2017","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1111\/exsy.12400","article-title":"A systematic review on deep learning architectures and applications","volume":"36","author":"Khamparia","year":"2019","journal-title":"Expert Syst."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Csillik, O., Cherbini, J., Johnson, R., Lyons, A., and Kelly, M. (2018). Identification of Citrus Trees from Unmanned Aerial Vehicle Imagery Using Convolutional Neural Networks. Drones, 2.","DOI":"10.3390\/drones2040039"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"349","DOI":"10.5194\/isprs-archives-XLII-2-W13-349-2019","article-title":"Crop row detection procedure using low-cost uav imagery system","volume":"42","author":"Hassanein","year":"2019","journal-title":"Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"5973","DOI":"10.1080\/01431161.2019.1584929","article-title":"Comparison of machine learning algorithms for classification of LiDAR points for characterization of canola canopy structure","volume":"40","author":"Wu","year":"2019","journal-title":"Int. J. Remote Sens."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Kitano, B.T., Mendes, C.C.T., Geus, A.R., Oliveira, H.C., and Souza, J.R. (2019). Corn Plant Counting Using Deep Learning and UAV Images. IEEE Geosci. Remote Sens. Lett., 1\u20135.","DOI":"10.1109\/LGRS.2019.2930549"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Dian Bah, M., Hafiane, A., and Canals, R. (2018). Deep learning with unsupervised data labeling for weed detection in line crops in UAV images. Remote Sens., 10.","DOI":"10.20944\/preprints201809.0088.v1"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Legg, M., and Bradley, S. (2019). Ultrasonic Arrays for Remote Sensing of Pasture Biomass. Remote Sens., 12.","DOI":"10.3390\/rs12010111"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Loggenberg, K., Strever, A., Greyling, B., and Poona, N. (2018). Modelling water stress in a Shiraz vineyard using hyperspectral imaging and machine learning. Remote Sens., 10.","DOI":"10.3390\/rs10020202"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"876","DOI":"10.1109\/JSTARS.2018.2793849","article-title":"Automatic Tobacco Plant Detection in UAV Images via Deep Neural Networks","volume":"11","author":"Fan","year":"2018","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_24","first-page":"79","article-title":"Combining UAV-based plant height from crop surface models, visible, and near infrared vegetation indices for biomass monitoring in barley","volume":"39","author":"Bendig","year":"2015","journal-title":"Int. J. Appl. Earth Obs. Geoinf."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"840","DOI":"10.1007\/s11119-018-9560-y","article-title":"Onion biomass monitoring using UAV-based RGB imaging","volume":"19","author":"Ballesteros","year":"2018","journal-title":"Precis. Agric."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Batistoti, J., Marcato Junior, J., \u00cdtavo, L., Matsubara, E., Gomes, E., Oliveira, B., Souza, M., Siqueira, H., Salgado Filho, G., and Akiyama, T. (2019). Estimating Pasture Biomass and Canopy Height in Brazilian Savanna Using UAV Photogrammetry. Remote Sens., 11.","DOI":"10.3390\/rs11202447"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"N\u00e4si, R., Viljanen, N., Kaivosoja, J., Alhonoja, K., Hakala, T., Markelin, L., and Honkavaara, E. (2018). Estimating Biomass and Nitrogen Amount of Barley and Grass Using UAV and Aircraft Based Spectral and Photogrammetric 3D Features. Remote Sens., 10.","DOI":"10.3390\/rs10071082"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"161","DOI":"10.1016\/j.isprsjprs.2020.02.013","article-title":"Above-ground biomass estimation and yield prediction in potato by using UAV-based RGB and hyperspectral imaging","volume":"162","author":"Li","year":"2020","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"70","DOI":"10.1016\/j.compag.2018.02.016","article-title":"Deep learning in agriculture: A survey","volume":"147","author":"Kamilaris","year":"2018","journal-title":"Comput. Electron. Agric."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"117","DOI":"10.1016\/j.eja.2018.12.004","article-title":"Estimating above ground biomass of winter wheat at early growth stages using digital images and deep convolutional neural network","volume":"103","author":"Ma","year":"2019","journal-title":"Eur. J. Agron."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"1132","DOI":"10.1071\/CP13319","article-title":"The value of improved pastures to Brazilian beef production","volume":"65","author":"Jank","year":"2014","journal-title":"Crop Pasture Sci."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1109\/MCSE.2007.55","article-title":"Matplotlib: A 2D graphics environment","volume":"9","author":"Hunter","year":"2007","journal-title":"Comput. Sci. Eng."},{"key":"ref_33","unstructured":"Gillies, S., Ward, B., and Petersen, A.S. (2020, August 17). Rasterio: Geospatial Raster I\/O for Python Programmers. Available online: https:\/\/github.com\/mapbox\/rasterio."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1109\/MCSE.2007.58","article-title":"Python for scientific computing","volume":"9","author":"Oliphant","year":"2007","journal-title":"Comput. Sci. Eng."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","article-title":"Deep learning","volume":"521","author":"LeCun","year":"2015","journal-title":"Nature"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"14","DOI":"10.1016\/j.knosys.2015.01.010","article-title":"Transfer learning using computational intelligence: A survey","volume":"80","author":"Lu","year":"2015","journal-title":"Knowl.-Based Syst."},{"key":"ref_37","unstructured":"Zhang, C., Bengio, S., Hardt, M., Recht, B., and Vinyals, O. (2016). Understanding deep learning requires rethinking generalization. arXiv."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1186\/s40537-019-0197-0","article-title":"A survey on image data augmentation for deep learning","volume":"6","author":"Shorten","year":"2019","journal-title":"J. Big Data"},{"key":"ref_39","unstructured":"Krizhevsky, A., Sutskever, I., and Hinton, G.E. (2020, August 10). Imagenet Classification with Deep Convolutional Neural Networks. Available online: http:\/\/papers.nips.cc\/paper\/4824-imagenet-classification-with-deep-convolutional-neural-networ."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_41","unstructured":"Simonyan, K., and Zisserman, A. (2014). Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv."},{"key":"ref_42","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., and Antiga, L. (2019, January 8\u201314). PyTorch: An imperative style, high-performance deep learning library. Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_43","unstructured":"Kingma, D.P., and Ba, J. (2014). Adam: A method for stochastic optimization. arXiv."},{"key":"ref_44","first-page":"781","article-title":"An empirical comparison of pattern recognition, neural nets, and machine learning classification methods","volume":"89","author":"Weiss","year":"1989","journal-title":"IJCAI"},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Blum, A., Kalai, A., and Langford, J. (1999, January 7\u20139). Beating the hold-out: Bounds for k-fold and progressive cross-validation. Proceedings of the Twelfth Annual Conference on Computational Learning Theory, Santa Cruz, CA, USA.","DOI":"10.1145\/307400.307439"},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"3395","DOI":"10.1016\/j.patcog.2013.06.014","article-title":"ROC curves for regression","volume":"46","year":"2013","journal-title":"Pattern Recognit."},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"267","DOI":"10.1007\/BF02289263","article-title":"Who belongs in the family","volume":"18","author":"Thorndike","year":"1953","journal-title":"Psychometrika"},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., and Torralba, A. (2016, January 27\u201330). Learning deep features for discriminative localization. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.319"},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017, January 22\u201329). Grad-cam: Visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Viljanen, N., Honkavaara, E., N\u00e4si, R., Hakala, T., Niemel\u00e4inen, O., and Kaivosoja, J. (2018). A novel machine learning method for estimating biomass of grass swards using a photogrammetric canopy height model, images and vegetation indices captured by a drone. Agriculture, 8.","DOI":"10.3390\/agriculture8050070"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Moeckel, T., Safari, H., Reddersen, B., Fricke, T., and Wachendorf, M. (2017). Fusion of ultrasonic and spectral sensor data for improving the estimation of biomass in grasslands with heterogeneous sward structure. Remote Sens., 9.","DOI":"10.3390\/rs9010098"},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1111\/gfs.12312","article-title":"Remote sensing as a tool to assess botanical composition, structure, quantity and quality of temperate grasslands","volume":"73","author":"Wachendorf","year":"2018","journal-title":"Grass Forage Sci."},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"10027","DOI":"10.3390\/s130810027","article-title":"Spectroscopic determination of aboveground biomass in grasslands using spectral transformations, support vector machine and partial least squares regression","volume":"13","author":"Marabel","year":"2013","journal-title":"Sensors"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/17\/4802\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T10:06:38Z","timestamp":1760177198000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/17\/4802"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8,26]]},"references-count":53,"journal-issue":{"issue":"17","published-online":{"date-parts":[[2020,9]]}},"alternative-id":["s20174802"],"URL":"https:\/\/doi.org\/10.3390\/s20174802","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,8,26]]}}}