{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,17]],"date-time":"2026-04-17T03:21:44Z","timestamp":1776396104208,"version":"3.51.2"},"reference-count":55,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2022,1,26]],"date-time":"2022-01-26T00:00:00Z","timestamp":1643155200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Monitoring crops and weeds is a major challenge in agriculture and food production today. Weeds compete directly with crops for moisture, nutrients, and sunlight. They therefore have a significant negative impact on crop yield if not sufficiently controlled. Weed detection and mapping is an essential step in weed control. Many existing research studies recognize the importance of remote sensing systems and machine learning algorithms in weed management. Deep learning approaches have shown good performance in many agriculture-related remote sensing tasks, such as plant classification, disease detection, etc. However, despite the success of these approaches, they still face many challenges such as high computation cost, the need of large labelled datasets, intra-class discrimination (in growing phase weeds and crops share many attributes similarity as color, texture, and shape), etc. This paper aims to show that the attention-based deep network is a promising approach to address the forementioned problems, in the context of weeds and crops recognition with drone system. The specific objective of this study was to investigate visual transformers (ViT) and apply them to plant classification in Unmanned Aerial Vehicles (UAV) images. Data were collected using a high-resolution camera mounted on a UAV, which was deployed in beet, parsley and spinach fields. The acquired data were augmented to build larger dataset, since ViT requires large sample sets for better performance, we also adopted the transfer learning strategy. Experiments were set out to assess the effect of training and validation dataset size, as well as the effect of increasing the test set while reducing the training set. The results show that with a small labeled training dataset, the ViT models outperform state-of-the-art models such as EfficientNet and ResNet. The results of this study are promising and show the potential of ViT to be applied to a wide range of remote sensing image analysis tasks.<\/jats:p>","DOI":"10.3390\/rs14030592","type":"journal-article","created":{"date-parts":[[2022,1,27]],"date-time":"2022-01-27T04:49:51Z","timestamp":1643258991000},"page":"592","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":191,"title":["Transformer Neural Network for Weed and Crop Classification of High Resolution UAV Images"],"prefix":"10.3390","volume":"14","author":[{"given":"Reenul","family":"Reedha","sequence":"first","affiliation":[{"name":"INSA CVL, University of Orleans, PRISME Laboratory EA 4229, 18022 Bourges, France"}]},{"given":"Eric","family":"Dericquebourg","sequence":"additional","affiliation":[{"name":"INSA CVL, University of Orleans, PRISME Laboratory EA 4229, 18022 Bourges, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9100-7539","authenticated-orcid":false,"given":"Raphael","family":"Canals","sequence":"additional","affiliation":[{"name":"INSA CVL, University of Orleans, PRISME Laboratory EA 4229, 45067 Orleans, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3185-9996","authenticated-orcid":false,"given":"Adel","family":"Hafiane","sequence":"additional","affiliation":[{"name":"INSA CVL, University of Orleans, PRISME Laboratory EA 4229, 18022 Bourges, France"}]}],"member":"1968","published-online":{"date-parts":[[2022,1,26]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"107148","DOI":"10.1016\/j.comnet.2020.107148","article-title":"A compilation of UAV applications for precision agriculture","volume":"172","author":"Sarigiannidis","year":"2020","journal-title":"Comput. Netw."},{"key":"ref_2","first-page":"633641","article-title":"Effect of Herbicides on Living Organisms in The Ecosystem and Available Alternative Control Methods","volume":"10","author":"Ustuner","year":"2020","journal-title":"Int. J. Sci. Res. Publ. (IJSRP)"},{"key":"ref_3","first-page":"453","article-title":"Weed and its management: A major threats to crop economy","volume":"6","author":"Patel","year":"2016","journal-title":"J. Pharm. Sci. Biosci. Res."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1885","DOI":"10.1080\/03650340.2019.1579904","article-title":"Investigation of alternate herbicides for effective weed management in glyphosate-tolerant cotton","volume":"65","author":"Iqbal","year":"2019","journal-title":"Arch. Agron. Soil Sci."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Vrbni\u010danin, S., Pavlovi\u0107, D., and Bo\u017ei\u0107, D. (2017). Weed Resistance to Herbicides. Herbicide Resistance in Weeds and Crops, IntechOpen Limited.","DOI":"10.5772\/67979"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"226","DOI":"10.1016\/j.compag.2019.02.005","article-title":"A review on weed detection using ground-based machine vision and image processing techniques","volume":"158","author":"Wang","year":"2019","journal-title":"Comput. Electron. Agric."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"322","DOI":"10.1002\/rob.21938","article-title":"Robotic weed control using automated weed and crop classification","volume":"37","author":"Wu","year":"2020","journal-title":"J. Field Robot."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"106273","DOI":"10.1016\/j.compag.2021.106273","article-title":"Computer vision-based citrus tree detection in a cultivated environment using UAV imagery","volume":"187","author":"Donmez","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"5189","DOI":"10.1109\/ACCESS.2019.2960873","article-title":"CRowNet: Deep Network for Crop Row Detection in UAV Images","volume":"8","author":"Bah","year":"2020","journal-title":"IEEE Access"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Huang, H., Deng, J., Lan, Y., Yang, A., Deng, X., and Zhang, L. (2018). A fully convolutional network for weed mapping of unmanned aerial vehicle (UAV) imagery. PLoS ONE, 13.","DOI":"10.1371\/journal.pone.0196302"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"3446","DOI":"10.1080\/01431161.2019.1706112","article-title":"Deep learning versus Object-based Image Analysis (OBIA) in weed mapping of UAV imagery","volume":"41","author":"Huang","year":"2020","journal-title":"Int. J. Remote Sens."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"1291","DOI":"10.1007\/s11119-020-09721-7","article-title":"Detection of Colchicum autumnale in drone images, using a machine-learning approach","volume":"21","author":"Petrich","year":"2020","journal-title":"Precis. Agric."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Puerto, A., Pedraza, C., Jamaica-Tenjo, D.A., and Osorio Delgado, A. (2020). A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images. AgriEngineering, 2.","DOI":"10.3390\/agriengineering2030032"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"551","DOI":"10.5194\/isprs-archives-XLII-3-W12-2020-551-2020","article-title":"Deep Convolutional Neural Networks for Weed Detection in Agricultural Crops using Optical Aerial Images","volume":"XLII-3\/W12-2020","author":"Ramirez","year":"2020","journal-title":"Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Patidar, S., Singh, U., Sharma, S. (2020, January 2\u20134). Weed Seedling Detection Using Mask Regional Convolutional Neural Network. In Proceedings of the 2020 International Conference on Electronics and Sustainable Communication Systems (ICESC), Coimbatore, India.","DOI":"10.1109\/ICESC48915.2020.9155701"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"588","DOI":"10.1109\/LRA.2017.2774979","article-title":"WeedNet: Dense Semantic Weed Classification Using Multispectral Images and MAV for Smart Farming","volume":"3","author":"Sa","year":"2017","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Sa, I., Popovic, M., Khanna, R., Chen, Z., Lottes, P., Liebisch, F., Nieto, J., Stachniss, C., Walter, A., and Siegwart, R. (2018). WeedMap: A Large-Scale Semantic Weed Mapping Framework Using Aerial Multispectral Imaging and Deep Neural Network for Precision Farming. Remote Sens., 10.","DOI":"10.3390\/rs10091423"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"314","DOI":"10.1016\/j.compag.2017.10.027","article-title":"Weed detection in soybean crops using ConvNets","volume":"143","author":"Pistori","year":"2017","journal-title":"Comput. Electron. Agric."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"41","DOI":"10.5194\/isprs-annals-IV-2-W3-41-2017","article-title":"Real-time blob-wise sugar beets vs weeds classification for monitoring fields using convolutional neural networks","volume":"IV-2\/W3","author":"Milioto","year":"2017","journal-title":"ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Sivakumar, A.N.V., Li, J., Scott, S., Psota, E., Jhala, A.J., Luck, J.D., and Shi, Y. (2020). Comparison of object detection and patch-based classification deep learning models on mid-to late-season weed detection in UAV imagery. Remote Sens., 12.","DOI":"10.3390\/rs12132136"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"106370","DOI":"10.1016\/j.compag.2021.106370","article-title":"Semantic segmentation model of cotton roots in-situ image based on attention mechanism","volume":"189","author":"Kang","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"105446","DOI":"10.1016\/j.compag.2020.105446","article-title":"Vine disease detection in uav multispectral images with deep learning segmentation approach","volume":"174","author":"Kerkech","year":"2020","journal-title":"Comput. Electron. Agric."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"184","DOI":"10.1016\/j.compag.2016.04.024","article-title":"A survey of image processing techniques for plant extraction and segmentation in the field","volume":"125","author":"Hamuda","year":"2016","journal-title":"Comput. Electron. Agric."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"e03685","DOI":"10.1016\/j.heliyon.2020.e03685","article-title":"An automatic visible-range video weed detection, segmentation and classification prototype in potato field","volume":"6","author":"Sabzi","year":"2020","journal-title":"Heliyon"},{"key":"ref_25","unstructured":"Saha, D. (2019). Development of Enhanced Weed Detection System with Adaptive Thresholding, K-Means and Support Vector Machine. Electronic Theses and Dissertations, South Dakota State University. Available online: https:\/\/openprairie.sdstate.edu\/cgi\/viewcontent.cgi?article=4399&context=etd."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"70","DOI":"10.1016\/j.compag.2018.02.016","article-title":"Deep learning in agriculture: A survey","volume":"147","author":"Kamilaris","year":"2018","journal-title":"Comput. Electron. Agric."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Wu, Z., Chen, Y., Zhao, B., Kang, X., and Ding, Y. (2021). Review of Weed Detection Methods Based on Computer Vision. Sensors, 21.","DOI":"10.3390\/s21113647"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","article-title":"Deep Learning","volume":"521","author":"LeCun","year":"2015","journal-title":"Nature"},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long Short-term Memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Comput."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"106067","DOI":"10.1016\/j.compag.2021.106067","article-title":"A survey of deep learning techniques for weed detection from images","volume":"184","author":"Hasan","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"ref_31","first-page":"143","article-title":"Generalization and network design strategies","volume":"19","author":"Lecun","year":"1989","journal-title":"Connect. Perspect."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"ImageNet Classification with Deep Convolutional Neural Networks","volume":"60","author":"Krizhevsky","year":"2012","journal-title":"Commun. ACM"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_34","unstructured":"Lecun, Y., Boser, B., Denker, J., Henderson, D., Howard, R., Hubbard, W., and Jackel, L. (1990). Handwritten digit recognition with a back-propagation network. Advances in Neural Information Processing Systems 2, Morgan Kaufmann Publishers Inc."},{"key":"ref_35","unstructured":"Nkemelu, D., Omeiza, D., and Lubalo, N. (2018). Deep Convolutional Neural Network for Plant Seedlings Classification. arXiv."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"50","DOI":"10.1016\/j.biosystemseng.2018.06.017","article-title":"Transfer learning for the classification of sugar beet and volunteer potato under field conditions","volume":"174","author":"Suh","year":"2018","journal-title":"Biosyst. Eng."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Dian Bah, M., Hafiane, A., and Canals, R. (2018). Deep learning with unsupervised data labeling for weed detection in line crops in UAV images. Remote Sens., 10.","DOI":"10.20944\/preprints201809.0088.v1"},{"key":"ref_38","unstructured":"Simonyan, K., and Zisserman, A. (2015). Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv."},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S.E., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2015, January 7\u201312). Going Deeper with Convolutions. Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. (2016, January 27\u201330). Rethinking the Inception Architecture for Computer Vision. Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.308"},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Hu, D. (2019, January 5\u20136). An Introductory Survey on Attention Mechanisms in NLP Problems. Proceedings of the IntelliSys, London, UK.","DOI":"10.1007\/978-3-030-29513-4_31"},{"key":"ref_42","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., and Polosukhin, I. (2017, January 4\u20139). Attention is all you need. Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. Available online: http:\/\/papers.neurips.cc\/paper\/7181-attention-is-all-you-need.pdf."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Khan, S., Naseer, M., Hayat, M., Zamir, S.W., Khan, F.S., and Shah, M. (2021). Transformers in Vision: A Survey. arXiv.","DOI":"10.1145\/3505244"},{"key":"ref_44","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv."},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"1345","DOI":"10.1109\/TKDE.2009.191","article-title":"A Survey on Transfer Learning","volume":"22","author":"Pan","year":"2010","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Shlens, J., and Le, Q.V. (2020, January 14\u201319). Randaugment: Practical automated data augmentation with a reduced search space. Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA.","DOI":"10.1109\/CVPRW50498.2020.00359"},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"48","DOI":"10.1016\/j.neucom.2021.03.091","article-title":"A review on the attention mechanism of deep learning","volume":"452","author":"Niu","year":"2021","journal-title":"Neurocomputing"},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Cheng, J., Dong, L., and Lapata, M. (2016, January 1\u20135). Long Short-Term Memory-Networks for Machine Reading. Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA.","DOI":"10.18653\/v1\/D16-1053"},{"key":"ref_49","unstructured":"Zhang, H., Goodfellow, I., Metaxas, D., and Odena, A. (2019, January 10\u201315). Self-attention generative adversarial networks. Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA."},{"key":"ref_50","unstructured":"Tan, M., and Le, Q.V. (2020). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv."},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"40","DOI":"10.1214\/09-SS054","article-title":"A survey of cross-validation procedures for model selection","volume":"4","author":"Arlot","year":"2010","journal-title":"Stat. Surv."},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"3301","DOI":"10.1093\/bioinformatics\/bti499","article-title":"Prediction error estimation: A comparison of resampling methods","volume":"21","author":"Molinaro","year":"2005","journal-title":"Bioinformatics"},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"168","DOI":"10.1016\/j.aci.2018.08.003","article-title":"Classification assessment methods","volume":"17","author":"Tharwat","year":"2020","journal-title":"Appl. Comput. Inform."},{"key":"ref_54","doi-asserted-by":"crossref","first-page":"427","DOI":"10.1016\/j.ipm.2009.03.002","article-title":"A systematic analysis of performance measures for classification tasks","volume":"45","author":"Sokolova","year":"2009","journal-title":"Inf. Process. Manag."},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"159","DOI":"10.1007\/978-3-540-72665-4_14","article-title":"Performance Measures in Classification of Human Communications","volume":"4509","author":"Sokolova","year":"2007","journal-title":"Advances in Artificial Intelligence"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/3\/592\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T22:08:19Z","timestamp":1760134099000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/3\/592"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,1,26]]},"references-count":55,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2022,2]]}},"alternative-id":["rs14030592"],"URL":"https:\/\/doi.org\/10.3390\/rs14030592","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,1,26]]}}}