{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T21:54:42Z","timestamp":1769723682953,"version":"3.49.0"},"reference-count":66,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2024,6,7]],"date-time":"2024-06-07T00:00:00Z","timestamp":1717718400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>An important consideration for UAV-based (unmanned aerial vehicle) object detection in the natural environment is vegetation height and foliar cover, which can visually obscure the items a machine learning model is trained to detect. Hence, the accuracy of aerial detection of objects such as surface landmines and UXO (unexploded ordnance) is highly dependent on the height and density of vegetation in a given area. In this study, we develop a model that estimates the detection accuracy (recall) of a YOLOv8 object\u2019s detection implementation as a function of occlusion due to vegetation coverage. To solve this function, we developed an algorithm to extract vegetation height and coverage of the UAV imagery from a digital surface model generated using structure-from-motion (SfM) photogrammetry. We find the relationship between recall and percent occlusion is well modeled by a sigmoid function using the PFM-1 landmine test case. Applying the sigmoid recall-occlusion relationship in conjunction with our vegetation cover algorithm to solve for percent occlusion, we mapped the uncertainty in detection rate due to vegetation in UAV-based SfM orthomosaics in eight different minefield environments. This methodology and model have significant implications for determining the optimal location and time of year for UAV-based object detection tasks and quantifying the uncertainty of deep learning object detection models in the natural environment.<\/jats:p>","DOI":"10.3390\/rs16122046","type":"journal-article","created":{"date-parts":[[2024,6,7]],"date-time":"2024-06-07T08:05:17Z","timestamp":1717747517000},"page":"2046","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Modeling the Effect of Vegetation Coverage on Unmanned Aerial Vehicles-Based Object Detection: A Study in the Minefield Environment"],"prefix":"10.3390","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7843-3939","authenticated-orcid":false,"given":"Jasper","family":"Baur","sequence":"first","affiliation":[{"name":"Demining Research Community, New York, NY 10027, USA"},{"name":"Lamont Doherty Earth Observatory, Columbia University, New York, NY 10027, USA"}]},{"given":"Kyle","family":"Dewey","sequence":"additional","affiliation":[{"name":"Demining Research Community, New York, NY 10027, USA"}]},{"given":"Gabriel","family":"Steinberg","sequence":"additional","affiliation":[{"name":"Demining Research Community, New York, NY 10027, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4137-547X","authenticated-orcid":false,"given":"Frank O.","family":"Nitsche","sequence":"additional","affiliation":[{"name":"Lamont Doherty Earth Observatory, Columbia University, New York, NY 10027, USA"}]}],"member":"1968","published-online":{"date-parts":[[2024,6,7]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., and Fei-Fei, L. (2009, January 20\u201325). Imagenet: A large-scale hierarchical image database. Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref_2","unstructured":"Krizhevsky, A., Sutskever, I., and Hinton, G.E. (2012, January 3\u20136). Imagenet classification with deep convolutional neural networks. Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA."},{"key":"ref_3","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (July, January 26). You only look once: Unified, real-time object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015, January 7\u201313). Fast r-cnn. Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile.","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref_5","unstructured":"Ren, S., He, K., Girshick, R., and Sun, J. (2015, January 7\u201312). Faster r-cnn: Towards real-time object detection with region proposal networks. Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016, January 11\u201314). Ssd: Single shot multibox detector. Proceedings of the Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands. Proceedings, Part I 14.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_7","first-page":"012022","article-title":"A review on the use of drones for precision agriculture","volume":"Volume 275","author":"Daponte","year":"2019","journal-title":"IOP Conference Series: Earth and Environmental Science"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"67","DOI":"10.30909\/vol.03.01.67114","article-title":"Volcanological applications of unoccupied aircraft systems (UAS): Developments, strategies, and future challenges","volume":"3","author":"James","year":"2020","journal-title":"Volcanica"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s12518-013-0120-x","article-title":"UAV for 3D mapping applications: A review","volume":"6","author":"Nex","year":"2014","journal-title":"Appl. Geomat."},{"key":"ref_10","unstructured":"Amoukteh, A., Janda, J., and Vincent, J. (2024, January 10). Drones Go to Work. BCG Global. Available online: https:\/\/www.bcg.com\/publications\/2017\/engineered-products-infrastructure-machinery-components-drones-go-work."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"300","DOI":"10.1016\/j.geomorph.2012.08.021","article-title":"\u2018Structure-from-Motion\u2019photogrammetry: A low-cost, effective tool for geoscience applications","volume":"179","author":"Westoby","year":"2012","journal-title":"Geomorphology"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Du, D., Qi, Y., Yu, H., Yang, Y., Duan, K., Li, G., Zhang, W., Huang, Q., and Tian, Q. (2018, January 8\u201314). The unmanned aerial vehicle benchmark: Object detection and tracking. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01249-6_23"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"104046","DOI":"10.1016\/j.imavis.2020.104046","article-title":"Deep learning-based object detection in low-altitude UAV datasets: A survey","volume":"104","author":"Mittal","year":"2020","journal-title":"Image Vis. Comput."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Cao, Y., He, Z., Wang, L., Wang, W., Yuan, Y., Zhang, D., Zhang, J., Zhu, P., Van Gool, L., and Han, J. (2021, January 11\u201317). VisDrone-DET2021: The vision meets drone object detection challenge results. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, BC, Canada.","DOI":"10.1109\/ICCVW54120.2021.00319"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Pham, M.T., Courtrai, L., Friguet, C., Lef\u00e8vre, S., and Baussard, A. (2020). YOLO-Fine: One-stage detector of small objects under various backgrounds in remote sensing images. Remote Sens., 12.","DOI":"10.3390\/rs12152501"},{"key":"ref_16","first-page":"102085","article-title":"Evaluating techniques for mapping island vegetation from unmanned aerial vehicle (UAV) images: Pixel classification, visual interpretation and machine learning approaches","volume":"89","author":"Hamylton","year":"2020","journal-title":"Int. J. Appl. Earth Obs. Geoinf."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Zangerl, U., Haselberger, S., and Kraushaar, S. (2022). Classifying Sparse Vegetation in a Proglacial Valley Using UAV Imagery and Random Forest Algorithm. Remote Sens., 14.","DOI":"10.3390\/rs14194919"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Baur, J., Steinberg, G., Nikulin, A., Chiu, K., and de Smet, T.S. (2020). Applying deep learning to automate UAV-based detection of scatterable landmines. Remote Sens., 12.","DOI":"10.3390\/rs12050859"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"21657","DOI":"10.1007\/s00521-022-07819-9","article-title":"A comprehensive review on landmine detection using deep learning techniques in 5G environment: Open issues and challenges","volume":"34","author":"Barnawi","year":"2022","journal-title":"Neural Comput. Appl."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Baji\u0107, M., and Poto\u010dnik, B. (2023). UAV Thermal Imaging for Unexploded Ordnance Detection by Using Deep Learning. Remote Sens., 15.","DOI":"10.3390\/rs15040967"},{"key":"ref_21","first-page":"9","article-title":"Computer Vision Detection of Explosive Ordnance: A High-Performance 9N235\/9N210 Cluster Submunition Detector","volume":"27","author":"Harvey","year":"2023","journal-title":"J. Conv. Weapons Destr."},{"key":"ref_22","first-page":"29","article-title":"How to implement drones and machine learning to reduce time, costs, and dangers associated with landmine detection","volume":"25","author":"Baur","year":"2021","journal-title":"J. Conv. Weapons Destr."},{"key":"ref_23","unstructured":"Coulloudon, B., Eshelman, K., Gianola, J., Habich, N., Hughes, L., Johnson, C., Pellant, M., Podborny, P., Rasmussen, A., and Robles, B. (1999). Sampling vegetation attributes, BLM Technical Reference."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Dong, Z., Lv, P., Zhang, Z., Qian, G., and Luo, W. (2012). Aeolian transport in the field: A comparison of the effects of different surface treatments. J. Geophys. Res. Atmos., 117.","DOI":"10.1029\/2012JD017538"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"516","DOI":"10.1007\/s13280-016-0770-0","article-title":"Changing Arctic snow cover: A review of recent developments and assessment of future needs for observations, modelling, and impacts","volume":"45","author":"Bokhorst","year":"2016","journal-title":"Ambio"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Saleh, K., Sz\u00e9n\u00e1si, S., and V\u00e1mossy, Z. (2021, January 21\u201323). Occlusion handling in generic object detection: A review. Proceedings of the 2021 IEEE 19th World Symposium on Applied Machine Intelligence and Informatics (SAMI), Herl\u2019any, Slovakia.","DOI":"10.1109\/SAMI50585.2021.9378657"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Dalborgo, V., Murari, T.B., Madureira, V.S., Moraes, J.G.L., Bezerra, V.M.O., Santos, F.Q., Silva, A., and Monteiro, R.L. (2023). Traffic Sign Recognition with Deep Learning: Vegetation Occlusion Detection in Brazilian Environments. Sensors, 23.","DOI":"10.3390\/s23135919"},{"key":"ref_28","unstructured":"ICBL-CMC (2023). International Campaign to Ban Landmines, Landmine Monitor 2023, ICBL-CMC."},{"key":"ref_29","unstructured":"GICHD (2023). Difficult Terrain in Mine Action, International Center for Humanitarian Demining."},{"key":"ref_30","first-page":"7","article-title":"Inspiring the Next Generation of Humanitarian Mine Action Researchers","volume":"27","author":"Tuohy","year":"2023","journal-title":"J. Conv. Weapons Destr."},{"key":"ref_31","unstructured":"National Mine Action Authority, and GICHD (2024, January 10). Explosive Ordnance Risk Education Interactive Map. ArcGIS Web Application. Available online: https:\/\/ua.imsma.org\/portal\/apps\/webappviewer\/index.html?id=92c5f2e0fa794acf95fefb20eebdecae."},{"key":"ref_32","unstructured":"ICBL-CMC (2023, December 05). Cluster Munition Coalition. Cluster Munition Monitor 2023. Available online: www.the-monitor.org."},{"key":"ref_33","unstructured":"Jean-Pierre, K., and Sullivan, J. (2023). Press Briefing by Press Secretary Karine Jean-Pierre and National Security Advisor Jake Sullivan, White House."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"1175","DOI":"10.1080\/01431161.2013.876120","article-title":"Mapping vegetation morphology types in a dry savanna ecosystem: Integrating hierarchical object-based image analysis with Random Forest","volume":"35","author":"Mishra","year":"2014","journal-title":"Int. J. Remote Sens."},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Resop, J.P., Lehmann, L., and Hession, W.C. (2021). Quantifying the spatial variability of annual and seasonal changes in riverscape vegetation using drone laser scanning. Drones, 5.","DOI":"10.3390\/drones5030091"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"393","DOI":"10.1007\/s10682-012-9601-3","article-title":"Functional traits of grasses growing in open and shaded habitats","volume":"27","author":"Cayssials","year":"2013","journal-title":"Evol. Ecol."},{"key":"ref_37","unstructured":"Jocher, G., Chaurasia, A., and Qiu, J. (2024, January 10). Ultralytics YOLOv8, Available online: https:\/\/github.com\/ultralytics\/ultralytics."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"195","DOI":"10.1016\/j.geomorph.2016.11.009","article-title":"An evaluation of the effectiveness of low-cost UAVs and structure from motion for geomorphic change detection","volume":"278","author":"Cook","year":"2017","journal-title":"Geomorphology"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1007\/s40725-019-00094-3","article-title":"Structure from motion photogrammetry in forestry: A review","volume":"5","author":"Iglhaut","year":"2019","journal-title":"Curr. For. Rep."},{"key":"ref_40","unstructured":"Pix4D (2020). How to Verify That There Is Enough Overlap between the Images\u2014Pix4D Mapper, Pix4D Support."},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"2002","DOI":"10.3389\/fpls.2017.02002","article-title":"High-throughput phenotyping of plant height: Comparing unmanned aerial vehicles and ground LiDAR estimates","volume":"8","author":"Madec","year":"2017","journal-title":"Front. Plant Sci."},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"886804","DOI":"10.3389\/fpls.2022.886804","article-title":"Comparison of Remote Sensing Methods for Plant Heights in Agricultural Fields Using Unmanned Aerial Vehicle-Based Structure from Motion","volume":"13","author":"Fujiwara","year":"2022","journal-title":"Front. Plant Sci."},{"key":"ref_43","first-page":"31","article-title":"Multitemporal field-based plant height estimation using 3D point clouds generated from small unmanned aerial systems high-resolution imagery","volume":"64","author":"Malambo","year":"2018","journal-title":"Int. J. Appl. Earth Obs. Geoinf."},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"671","DOI":"10.1038\/nmeth.2089","article-title":"NIH Image to ImageJ: 25 years of image analysis","volume":"9","author":"Schneider","year":"2012","journal-title":"Nat. Methods"},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"922","DOI":"10.3390\/f4040922","article-title":"A photogrammetric workflow for the creation of a forest canopy height model from small unmanned aerial system imagery","volume":"4","author":"Lisein","year":"2013","journal-title":"Forests"},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"DiGiacomo, A.E., Bird, C.N., Pan, V.G., Dobroski, K., Atkins-Davis, C., Johnston, D.W., and Ridge, J.T. (2020). Modeling salt marsh vegetation height using unoccupied aircraft systems and structure from motion. Remote Sens., 12.","DOI":"10.3390\/rs12142333"},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"452","DOI":"10.1080\/1343943X.2020.1766362","article-title":"Field phenotyping of plant height in an upland rice field in Laos using low-cost small unmanned aerial vehicles (UAVs)","volume":"23","author":"Kawamura","year":"2020","journal-title":"Plant Prod. Sci."},{"key":"ref_48","unstructured":"Zhong, Z., Zheng, L., Kang, G., Li, S., and Yang, Y. (2020, January 7\u201312). Random erasing data augmentation. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s40537-019-0197-0","article-title":"A survey on image data augmentation for deep learning","volume":"6","author":"Shorten","year":"2019","journal-title":"J. Big Data"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Ke, L., Tai, Y.W., and Tang, C.K. (2021, January 20\u201325). Deep occlusion-aware instance segmentation with overlapping bilayers. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.00401"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Yuan, X., Kortylewski, A., Sun, Y., and Yuille, A. (2021, January 20\u201325). Robust instance segmentation through reasoning about multi-object occlusion. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01099"},{"key":"ref_52","doi-asserted-by":"crossref","unstructured":"Gao, T., Packer, B., and Koller, D. (2011, January 20\u201325). A segmentation-aware object detection model with occlusion handling. Proceedings of the CVPR 2011, Colorado Springs, CO, USA.","DOI":"10.1109\/CVPR.2011.5995623"},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"40","DOI":"10.1016\/j.isprsjprs.2016.12.009","article-title":"A survey of landmine detection using hyperspectral imaging","volume":"124","author":"Makki","year":"2017","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_54","doi-asserted-by":"crossref","unstructured":"Nikulin, A., De Smet, T.S., Baur, J., Frazer, W.D., and Abramowitz, J.C. (2018). Detection and identification of remnant PFM-1 \u2018Butterfly Mines\u2019 with a UAV-based thermal-imaging protocol. Remote Sens., 10.","DOI":"10.3390\/rs10111672"},{"key":"ref_55","doi-asserted-by":"crossref","unstructured":"Qiu, Z., Guo, H., Hu, J., Jiang, H., and Luo, C. (2023). Joint Fusion and Detection via Deep Learning in UAV-Borne Multispectral Sensing of Scatterable Landmine. Sensors, 23.","DOI":"10.3390\/s23125693"},{"key":"ref_56","doi-asserted-by":"crossref","first-page":"9341","DOI":"10.1109\/JSEN.2019.2925203","article-title":"Landmine detection using multispectral images","volume":"19","author":"Silva","year":"2019","journal-title":"IEEE Sens. J."},{"key":"ref_57","unstructured":"U.S. Department of Agriculture, Foreign Agricultural Service (2024, January 10). Ukraine Agricultural Production and Trade\u2014April 2022, Available online: https:\/\/www.fas.usda.gov\/sites\/default\/files\/2022-04\/Ukraine-Factsheet-April2022.pdf."},{"key":"ref_58","first-page":"286","article-title":"Light detection and ranging (LIDAR): An emerging tool for multiple resource inventory","volume":"103","author":"Reutebuch","year":"2005","journal-title":"J. For."},{"key":"ref_59","doi-asserted-by":"crossref","unstructured":"ten Harkel, J., Bartholomeus, H., and Kooistra, L. (2019). Biomass and crop height estimation of different crops using UAV-based LiDAR. Remote Sens., 12.","DOI":"10.3390\/rs12010017"},{"key":"ref_60","doi-asserted-by":"crossref","first-page":"2014","DOI":"10.1109\/TGRS.2008.2010490","article-title":"Separation of ground and low vegetation signatures in LiDAR measurements of salt-marsh environments","volume":"47","author":"Wang","year":"2009","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_61","doi-asserted-by":"crossref","unstructured":"Dhami, H., Yu, K., Xu, T., Zhu, Q., Dhakal, K., Friel, J., Li, S., and Tokekar, P. (2020\u201324, January 24). Crop height and plot estimation for phenotyping from unmanned aerial vehicles using 3D LiDAR. Proceedings of the 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA.","DOI":"10.1109\/IROS45743.2020.9341343"},{"key":"ref_62","doi-asserted-by":"crossref","first-page":"1337","DOI":"10.1109\/TPAMI.2003.1233909","article-title":"Detecting moving objects, ghosts, and shadows in video streams","volume":"25","author":"Cucchiara","year":"2003","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_63","unstructured":"Lee, J.T., Lim, K.T., and Chung, Y. (2016). Moving shadow detection from background image and deep learning. Image and Video Technology\u2013PSIVT 2015 Workshops: RV 2015, GPID 2013, VG 2015, EO4AS 2015, MCBMIIA 2015, and VSWS 2015, Auckland, New Zealand, 23\u201327 November 2015, Springer International Publishing. Revised Selected Papers 7."},{"key":"ref_64","doi-asserted-by":"crossref","first-page":"1222","DOI":"10.1016\/j.patcog.2006.09.017","article-title":"Shadow detection for moving objects based on texture analysis","volume":"40","author":"Leone","year":"2007","journal-title":"Pattern Recognit."},{"key":"ref_65","doi-asserted-by":"crossref","unstructured":"Pal, M., Palevi\u010dius, P., Landauskas, M., Orinait\u0117, U., Timofejeva, I., and Ragulskis, M. (2021). An overview of challenges associated with automatic detection of concrete cracks in the presence of shadows. Appl. Sci., 11.","DOI":"10.3390\/app112311396"},{"key":"ref_66","doi-asserted-by":"crossref","first-page":"63971","DOI":"10.1109\/ACCESS.2020.2984680","article-title":"A moving shadow elimination method based on fusion of multi-feature","volume":"8","author":"Zhang","year":"2020","journal-title":"IEEE Access"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/12\/2046\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T14:55:14Z","timestamp":1760108114000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/12\/2046"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,7]]},"references-count":66,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2024,6]]}},"alternative-id":["rs16122046"],"URL":"https:\/\/doi.org\/10.3390\/rs16122046","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,7]]}}}