{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,17]],"date-time":"2026-02-17T02:47:03Z","timestamp":1771296423283,"version":"3.50.1"},"reference-count":40,"publisher":"MDPI AG","issue":"11","license":[{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Maatalouskoneiden tutkimuss\u00e4\u00e4ti\u00f6 (Agricultural Machinery Research Foundation)"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>In the field of precision agriculture, weed detection models combined with selective spraying by ground or aerial robotics are seen as promising approaches for increasing yield harvests while simultaneously minimizing herbicide use. The available labeled training data are a major bottleneck for developing and applying supervised deep learning models, and more automated methods for labeled data generation are therefore needed. Our research aims to address this need by introducing a rule-based method for label data generation for perennial weeds. For this research, a dataset of a barley field was collected using an unmanned aerial vehicle (UAV) with a flight altitude of 10 m. A multispectral and a thermal camera were used for the data collection. The preprocessed dataset consists of multispectral and thermal orthomosaic images along with a canopy height model. The methodological part of this article introduces a proposed rule-based method for labeled data generation for perennial weeds based on the Normalized Difference Vegetation Index (NDVI), and this approach is further used to generate labels for the measured data. The preprocessed data combined with the generated labels was used to train U-net models. Three data combinations are used for the training and testing: multispectral, multispectral\u2013thermal and multispectral\u2013thermal\u2013canopy\u2013height\u2013model. This approach was used to evaluate whether additional data improve model performance. To evaluate the models on ground truth labels, they are tested on a manually annotated test dataset, which consists of 10% of the whole dataset. The tested models achieved an F1 score of 0.82\u20130.83 on the test dataset. This finding demonstrates that the rule-based labeling method generates valid labels for the perennial weed detection task. Furthermore, our study shows that data fusion improved the model slightly. The data combination of multispectral\u2013thermal\u2013canopy\u2013height\u2013model as input resulted in the best-performing model, with an F1 score of 0.835.<\/jats:p>","DOI":"10.3390\/rs15112877","type":"journal-article","created":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T03:06:00Z","timestamp":1685588760000},"page":"2877","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Weakly Supervised Perennial Weed Detection in a Barley Field"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-8194-5149","authenticated-orcid":false,"given":"Leon-Friedrich","family":"Thomas","sequence":"first","affiliation":[{"name":"Department of Agricultural Sciences, University of Helsinki, 00790 Helsinki, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2599-1894","authenticated-orcid":false,"given":"Mikael","family":"\u00c4n\u00e4kk\u00e4l\u00e4","sequence":"additional","affiliation":[{"name":"Department of Agricultural Sciences, University of Helsinki, 00790 Helsinki, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2175-7833","authenticated-orcid":false,"given":"Antti","family":"Lajunen","sequence":"additional","affiliation":[{"name":"Department of Agricultural Sciences, University of Helsinki, 00790 Helsinki, Finland"}]}],"member":"1968","published-online":{"date-parts":[[2023,6,1]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"359","DOI":"10.1023\/B:PRAG.0000040806.39604.aa","article-title":"Precision Agriculture and Sustainability","volume":"5","author":"Bongiovanni","year":"2004","journal-title":"Precis. Agric."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Monteiro, A., and Santos, S. (2022). Sustainable Approach to Weed Management: The Role of Precision Weed Management. Agronomy, 12.","DOI":"10.3390\/agronomy12010118"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1186\/s40538-021-00217-8","article-title":"Drone and sensor technology for sustainable weed management: A review","volume":"8","author":"Esposito","year":"2021","journal-title":"Chem. Biol. Technol. Agric."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"106067","DOI":"10.1016\/j.compag.2021.106067","article-title":"A survey of deep learning techniques for weed detection from images","volume":"184","author":"Hasan","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Wu, Z., Chen, Y., Zhao, B., Kang, X., and Ding, Y. (2021). Review of Weed Detection Methods Based on Computer Vision. Sensors, 21.","DOI":"10.3390\/s21113647"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"226","DOI":"10.1016\/j.compag.2019.02.005","article-title":"A review on weed detection using ground-based machine vision and image processing techniques","volume":"158","author":"Wang","year":"2019","journal-title":"Comput. Electron. Agric."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Bah, M., Hafiane, A., and Canals, R. (2018). Deep Learning with Unsupervised Data Labeling for Weed Detection in Line Crops in UAV Images. Remote Sens., 10.","DOI":"10.20944\/preprints201809.0088.v1"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"588","DOI":"10.1109\/LRA.2017.2774979","article-title":"weedNet: Dense Semantic Weed Classification Using Multispectral Images and MAV for Smart Farming","volume":"3","author":"Sa","year":"2018","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Sheikh, R., Milioto, A., Lottes, P., Stachniss, C., Bennewitz, M., and Schultz, T. (August, January 31). Gradient and Log-based Active Learning for Semantic Segmentation of Crop and Weed for Agricultural Robots. Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France.","DOI":"10.1109\/ICRA40945.2020.9196722"},{"key":"ref_10","unstructured":"Mishra, A., and Gautam, V. (2021, January 25\u201327). Weed Species Identification in Different Crops Using Precision Weed Management: A Review. Proceedings of the International Symposium on Intelligent Control, New Delhi, India."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Tsouros, D.C., Bibi, S., and Sarigiannidis, P.G. (2019). A Review on UAV-Based Applications for Precision Agriculture. Information, 10.","DOI":"10.3390\/info10110349"},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"102461","DOI":"10.1016\/j.jnca.2019.102461","article-title":"A survey of unmanned aerial sensing solutions in precision agriculture","volume":"148","author":"Mukherjee","year":"2019","journal-title":"J. Netw. Comput. Appl."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"482","DOI":"10.1007\/s11119-022-09954-8","article-title":"Early\/late fusion structures with optimized feature selection for weed detection using visible and thermal images of paddy fields","volume":"24","author":"Zamani","year":"2023","journal-title":"Precis. Agric."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Shahbazi, N., Ashworth, M.B., Callow, J.N., Mian, A., Beckie, H.J., Speidel, S., Nicholls, E., and Flower, K.C. (2021). Assessing the Capability and Potential of LiDAR for Weed Detection. Sensors, 21.","DOI":"10.3390\/s21072328"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Torres-S\u00e1nchez, J., Mesas-Carrascosa, F.J., Jim\u00e9nez-Brenes, F.M., de Castro, A.I., and L\u00f3pez-Granados, F. (2021). Early Detection of Broad-Leaved and Grass Weeds in Wide Row Crops Using Artificial Neural Networks and UAV Imagery. Agronomy, 11.","DOI":"10.3390\/agronomy11040749"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"61","DOI":"10.54480\/slrm.v2i2.21","article-title":"Weed detection using machine learning: A systematic literature review","volume":"2","author":"Abubakar","year":"2021","journal-title":"Syst. Lit. Rev. Meta-Anal. J."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Rakhmatulin, I., Kamilaris, A., and Andreasen, C. (2021). Deep Neural Networks to Detect Weeds from Crops in Agricultural Environments in Real-Time: A Review. Remote Sens., 13.","DOI":"10.2139\/ssrn.3959386"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"842","DOI":"10.1017\/S2040470017000206","article-title":"RoboWeedSupport\u2014Detection of weed locations in leaf occluded cereal crops using a fully convolutional neural network","volume":"8","author":"Dyrmann","year":"2017","journal-title":"Adv. Anim. Biosci."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Reedha, R., Dericquebourg, E., Canals, R., and Hafiane, A. (2022). Transformer Neural Network for Weed and Crop Classification of High Resolution UAV Images. Remote Sens., 14.","DOI":"10.3390\/rs14030592"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"245","DOI":"10.2137\/145960611797471534","article-title":"A Composition of weed flora in spring cereals in Finland\u2014A fourth survey","volume":"20","author":"Salonen","year":"2011","journal-title":"Agric. Food Sci."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Sa, I., Popovi\u0107, M., Khanna, R., Chen, Z., Lottes, P., Liebisch, F., Nieto, J., Stachniss, C., Walter, A., and Siegwart, R. (2018). WeedMap: A Large-Scale Semantic Weed Mapping Framework Using Aerial Multispectral Imaging and Deep Neural Network for Precision Farming. Remote Sens., 10.","DOI":"10.3390\/rs10091423"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Ramirez, W., Achanccaray, P., Mendoza, L.A.F., and Pacheco, M.A.C. (2020, January 22\u201326). Deep Convolutional Neural Networks for Weed Detection in Agricultural Crops Using Optical Aerial Images. Proceedings of the 2020 IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS), Santiago, Chile.","DOI":"10.1109\/LAGIRS48042.2020.9165562"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Boyd, K., Eng, K.H., and Page, C.D. (2013, January 23\u201327). Erratum: Area under the Precision-Recall Curve: Point Estimates and Confidence Intervals. Proceedings of the Machine Learning and Knowledge Discovery in Databases, Prague, Czech Republic.","DOI":"10.1007\/978-3-642-40994-3_55"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"471","DOI":"10.3390\/agriengineering2030032","article-title":"A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images","volume":"2","author":"Osorio","year":"2020","journal-title":"AgriEngineering"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Eide, A., Koparan, C., Zhang, Y., Ostlie, M., Howatt, K., and Sun, X. (2021). UAV-Assisted Thermal Infrared and Multispectral Imaging of Weed Canopies for Glyphosate Resistance Detection. Remote Sens., 13.","DOI":"10.3390\/rs13224606"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"452","DOI":"10.1080\/1343943X.2020.1766362","article-title":"Field phenotyping of plant height in an upland rice field in Laos using low-cost small unmanned aerial vehicles (UAVs)","volume":"23","author":"Kawamura","year":"2020","journal-title":"Plant Prod. Sci."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Zisi, T., Alexandridis, T., Kaplanis, S., Navrozidis, I., Tamouridou, A.-A., Lagopodi, A., Moshou, D., and Polychronos, V. (2018). Incorporating Surface Elevation Information in UAV Multispectral Images for Mapping Weed Patches. J. Imaging, 4.","DOI":"10.3390\/jimaging4110132"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"198","DOI":"10.1080\/1343943X.2020.1829490","article-title":"Discriminating crops\/weeds in an upland rice field from UAV images with the SLIC-RF algorithm","volume":"24","author":"Kawamura","year":"2020","journal-title":"Plant Prod. Sci."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"16","DOI":"10.1016\/j.cropro.2017.03.006","article-title":"Control of perennial weeds in spring cereals through stubble cultivation and mouldboard ploughing during autumn or spring","volume":"98","author":"Mangerud","year":"2017","journal-title":"Crop Prot."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Messina, G., and Modica, G. (2020). Applications of UAV Thermal Imagery in Precision Agriculture: State of the Art and Future Research Outlook. Remote Sens., 12.","DOI":"10.3390\/rs12091491"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Maguire, M.S., Neale, C.M.U., and Woldt, W.E. (2021). Improving Accuracy of Unmanned Aerial System Thermal Infrared Remote Sensing for Use in Energy Balance Models in Agriculture Applications. Remote Sens., 13.","DOI":"10.3390\/rs13091635"},{"key":"ref_32","unstructured":"Finnish Meteorological Institute (2023, January 16). Relative Humidity and Air Temperature at Helsinki Malmi Lentokentt\u00e4. Available online: https:\/\/en.ilmatieteenlaitos.fi\/download-observations."},{"key":"ref_33","unstructured":"Rouse, J.W., Haas, R.H., Schell, J., and Deering, D. (1973). Monitoring the Vernal Advancement and Retrogradation (Green Wave Effect) of Natural Vegetation, Remote Sensing Center, Texas A&M University."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"31","DOI":"10.1109\/TMI.1983.4307610","article-title":"Comparison of interpolating methods for image resampling","volume":"2","author":"Parker","year":"1983","journal-title":"IEEE Trans. Med. Imaging"},{"key":"ref_35","first-page":"159","article-title":"Case Study: Cost-effective Weed Patch Detection by Multi-Spectral Camera Mounted on Unmanned Aerial Vehicle in the Buckwheat Field","volume":"64","author":"Kim","year":"2019","journal-title":"Korean J. Crop Sci."},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., and Brox, T. (2015). U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv.","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref_37","unstructured":"Comble, A.d.L., and Prepin, K. (2021, January 25\u201327). Efficient transfer learning for multi-channel convolutional neural networks. Proceedings of the 2021 17th International Conference on Machine Vision and Applications (MVA), Aichi, Japan."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1093\/oxfordjournals.pan.a004868","article-title":"Logistic Regression in Rare Events Data","volume":"9","author":"King","year":"2001","journal-title":"Political Anal."},{"key":"ref_39","unstructured":"Kingma, D., and Ba, J. (2015, January 7\u20139). Adam: A Method for Stochastic Optimization. Proceedings of the International Conference on Learning Representations, San Diego, CA, USA."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Rahman, M.A., and Wang, Y. (2016, January 12\u201314). Optimizing Intersection-Over-Union in Deep Neural Networks for Image Segmentation. Proceedings of the Advances in Visual Computing, Las Vegas, NV, USA.","DOI":"10.1007\/978-3-319-50835-1_22"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/11\/2877\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T19:46:54Z","timestamp":1760125614000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/11\/2877"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,1]]},"references-count":40,"journal-issue":{"issue":"11","published-online":{"date-parts":[[2023,6]]}},"alternative-id":["rs15112877"],"URL":"https:\/\/doi.org\/10.3390\/rs15112877","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,1]]}}}