{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T18:22:29Z","timestamp":1772302949428,"version":"3.50.1"},"reference-count":34,"publisher":"MDPI AG","issue":"14","license":[{"start":{"date-parts":[[2023,7,11]],"date-time":"2023-07-11T00:00:00Z","timestamp":1689033600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Laboratory of Lingnan Modern Agriculture 713 Science and Technology Guangdong Experimental Heyuan Branch Project 714","award":["DT20220010"],"award-info":[{"award-number":["DT20220010"]}]},{"name":"Laboratory of Lingnan Modern Agriculture 713 Science and Technology Guangdong Experimental Heyuan Branch Project 714","award":["2021XNYNYKJHZGJ032"],"award-info":[{"award-number":["2021XNYNYKJHZGJ032"]}]},{"name":"Laboratory of Lingnan Modern Agriculture 713 Science and Technology Guangdong Experimental Heyuan Branch Project 714","award":["2022KJ108"],"award-info":[{"award-number":["2022KJ108"]}]},{"name":"Laboratory of Lingnan Modern Agriculture 713 Science and Technology Guangdong Experimental Heyuan Branch Project 714","award":["32271997"],"award-info":[{"award-number":["32271997"]}]},{"name":"Laboratory of Lingnan Modern Agriculture 713 Science and Technology Guangdong Experimental Heyuan Branch Project 714","award":["pdjh2023a0074"],"award-info":[{"award-number":["pdjh2023a0074"]}]},{"name":"Laboratory of Lingnan Modern Agriculture 713 Science and Technology Guangdong Experimental Heyuan Branch Project 714","award":["pdjh2021b0077"],"award-info":[{"award-number":["pdjh2021b0077"]}]},{"name":"Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University","award":["DT20220010"],"award-info":[{"award-number":["DT20220010"]}]},{"name":"Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University","award":["2021XNYNYKJHZGJ032"],"award-info":[{"award-number":["2021XNYNYKJHZGJ032"]}]},{"name":"Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University","award":["2022KJ108"],"award-info":[{"award-number":["2022KJ108"]}]},{"name":"Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University","award":["32271997"],"award-info":[{"award-number":["32271997"]}]},{"name":"Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University","award":["pdjh2023a0074"],"award-info":[{"award-number":["pdjh2023a0074"]}]},{"name":"Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University","award":["pdjh2021b0077"],"award-info":[{"award-number":["pdjh2021b0077"]}]},{"name":"the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China","award":["DT20220010"],"award-info":[{"award-number":["DT20220010"]}]},{"name":"the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China","award":["2021XNYNYKJHZGJ032"],"award-info":[{"award-number":["2021XNYNYKJHZGJ032"]}]},{"name":"the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China","award":["2022KJ108"],"award-info":[{"award-number":["2022KJ108"]}]},{"name":"the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China","award":["32271997"],"award-info":[{"award-number":["32271997"]}]},{"name":"the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China","award":["pdjh2023a0074"],"award-info":[{"award-number":["pdjh2023a0074"]}]},{"name":"the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China","award":["pdjh2021b0077"],"award-info":[{"award-number":["pdjh2021b0077"]}]},{"name":"National Natural Science Foundation of China","award":["DT20220010"],"award-info":[{"award-number":["DT20220010"]}]},{"name":"National Natural Science Foundation of China","award":["2021XNYNYKJHZGJ032"],"award-info":[{"award-number":["2021XNYNYKJHZGJ032"]}]},{"name":"National Natural Science Foundation of China","award":["2022KJ108"],"award-info":[{"award-number":["2022KJ108"]}]},{"name":"National Natural Science Foundation of China","award":["32271997"],"award-info":[{"award-number":["32271997"]}]},{"name":"National Natural Science Foundation of China","award":["pdjh2023a0074"],"award-info":[{"award-number":["pdjh2023a0074"]}]},{"name":"National Natural Science Foundation of China","award":["pdjh2021b0077"],"award-info":[{"award-number":["pdjh2021b0077"]}]},{"name":"the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (\u201cClimbing Program\u201d Special Fund), China","award":["DT20220010"],"award-info":[{"award-number":["DT20220010"]}]},{"name":"the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (\u201cClimbing Program\u201d Special Fund), China","award":["2021XNYNYKJHZGJ032"],"award-info":[{"award-number":["2021XNYNYKJHZGJ032"]}]},{"name":"the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (\u201cClimbing Program\u201d Special Fund), China","award":["2022KJ108"],"award-info":[{"award-number":["2022KJ108"]}]},{"name":"the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (\u201cClimbing Program\u201d Special Fund), China","award":["32271997"],"award-info":[{"award-number":["32271997"]}]},{"name":"the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (\u201cClimbing Program\u201d Special Fund), China","award":["pdjh2023a0074"],"award-info":[{"award-number":["pdjh2023a0074"]}]},{"name":"the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (\u201cClimbing Program\u201d Special Fund), China","award":["pdjh2021b0077"],"award-info":[{"award-number":["pdjh2021b0077"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Crop classification of large-scale agricultural land is crucial for crop monitoring and yield estimation. Hyperspectral image classification has proven to be an effective method for this task. Most current popular hyperspectral image classification methods are based on image classification, specifically on convolutional neural networks (CNNs) and recurrent neural networks (RNNs). In contrast, this paper focuses on methods based on semantic segmentation and proposes a new transformer-based approach called HyperSFormer for crop hyperspectral image classification. The key enhancement of the proposed method is the replacement of the encoder in SegFormer with an improved Swin Transformer while keeping the SegFormer decoder. The entire model adopts a simple and uniform transformer architecture. Additionally, the paper introduces the hyper patch embedding (HPE) module to extract spectral and local spatial information from the hyperspectral images, which enhances the effectiveness of the features used as input for the model. To ensure detailed model processing and achieve end-to-end hyperspectral image classification, the transpose padding upsample (TPU) module is proposed for the model\u2019s output. In order to address the problem of insufficient and imbalanced samples in hyperspectral image classification, the paper designs an adaptive min log sampling (AMLS) strategy and a loss function that incorporates dice loss and focal loss to assist model training. Experimental results using three public hyperspectral image datasets demonstrate the strong performance of HyperSFormer, particularly in the presence of imbalanced sample data, complex negative samples, and mixed sample classes. HyperSFormer outperforms state-of-the-art methods, including fast patch-free global learning (FPGA), a spectral\u2013spatial-dependent global learning framework (SSDGL), and SegFormer, by at least 2.7% in the mean intersection over union (mIoU). It also improves the overall accuracy and average accuracy values by at least 0.9% and 0.3%, respectively, and the kappa coefficient by at least 0.011. Furthermore, ablation experiments were conducted to determine the optimal hyperparameter and loss function settings for the proposed method, validating the rationality of these settings and the fusion loss function.<\/jats:p>","DOI":"10.3390\/rs15143491","type":"journal-article","created":{"date-parts":[[2023,7,12]],"date-time":"2023-07-12T01:01:41Z","timestamp":1689123701000},"page":"3491","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":31,"title":["HyperSFormer: A Transformer-Based End-to-End Hyperspectral Image Classification Method for Crop Classification"],"prefix":"10.3390","volume":"15","author":[{"given":"Jiaxing","family":"Xie","sequence":"first","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"},{"name":"Laboratory of Lingnan Modern Agriculture Science and Technology Guangdong Experimental Heyuan Branch, Heyuan 514000, China"},{"name":"Engineering Research Center for Monitoring Agricultural Information of Guangdong Province, Guangzhou 510642, China"}]},{"given":"Jiajun","family":"Hua","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"}]},{"given":"Shaonan","family":"Chen","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"}]},{"given":"Peiwen","family":"Wu","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2729-1165","authenticated-orcid":false,"given":"Peng","family":"Gao","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"}]},{"given":"Daozong","family":"Sun","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"},{"name":"Laboratory of Lingnan Modern Agriculture Science and Technology Guangdong Experimental Heyuan Branch, Heyuan 514000, China"},{"name":"Engineering Research Center for Monitoring Agricultural Information of Guangdong Province, Guangzhou 510642, China"}]},{"given":"Zhendong","family":"Lyu","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"}]},{"given":"Shilei","family":"Lyu","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"},{"name":"Laboratory of Lingnan Modern Agriculture Science and Technology Guangdong Experimental Heyuan Branch, Heyuan 514000, China"},{"name":"Engineering Research Center for Monitoring Agricultural Information of Guangdong Province, Guangzhou 510642, China"}]},{"given":"Xiuyun","family":"Xue","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"},{"name":"Laboratory of Lingnan Modern Agriculture Science and Technology Guangdong Experimental Heyuan Branch, Heyuan 514000, China"},{"name":"Engineering Research Center for Monitoring Agricultural Information of Guangdong Province, Guangzhou 510642, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6417-4646","authenticated-orcid":false,"given":"Jianqiang","family":"Lu","sequence":"additional","affiliation":[{"name":"College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China"},{"name":"Laboratory of Lingnan Modern Agriculture Science and Technology Guangdong Experimental Heyuan Branch, Heyuan 514000, China"},{"name":"Engineering Research Center for Monitoring Agricultural Information of Guangdong Province, Guangzhou 510642, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,7,11]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"2249","DOI":"10.1109\/JSTARS.2016.2639043","article-title":"Radar Remote Sensing of Agricultural Canopies: A Review","volume":"10","author":"McNairn","year":"2017","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1109\/MGRS.2016.2616418","article-title":"Advanced Spectral Classifiers for Hyperspectral Images: A Review","volume":"5","author":"Ghamisi","year":"2017","journal-title":"IEEE Geosci. Remote Sens. Mag."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"126241","DOI":"10.1016\/j.eja.2021.126241","article-title":"An Overview of Crop Nitrogen Status Assessment Using Hyperspectral Remote Sensing: Current Status and Perspectives","volume":"124","author":"Fu","year":"2021","journal-title":"Eur. J. Agron."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"11709","DOI":"10.1109\/TCYB.2021.3070577","article-title":"A Spectral-Spatial-Dependent Global Learning Framework for Insufficient and Imbalanced Hyperspectral Image Classification","volume":"52","author":"Zhu","year":"2021","journal-title":"IEEE Trans. Cybern."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Tinega, H.C., Chen, E., and Nyasaka, D.O. (2023). Improving Feature Learning in Remote Sensing Images Using an Integrated Deep Multi-Scale 3D\/2D Convolutional Network. Remote Sens., 15.","DOI":"10.3390\/rs15133270"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Padilla-Zepeda, E., Torres-Roman, D., and Mendez-Vazquez, A. (2023). A Semantic Segmentation Framework for Hyperspectral Imagery Based on Tucker Decomposition and 3DCNN Tested with Simulated Noisy Scenarios. Remote Sens., 15.","DOI":"10.3390\/rs15051399"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Liang, L., Zhang, S., Li, J., Plaza, A., and Cui, Z. (2023). Multi-Scale Spectral-Spatial Attention Network for Hyperspectral Image Classification Combining 2D Octave and 3D Convolutional Neural Networks. Remote Sens., 15.","DOI":"10.3390\/rs15071758"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"4237","DOI":"10.1109\/TGRS.2019.2961947","article-title":"Spatial\u2013Spectral Feature Extraction via Deep ConvLSTM Neural Networks for Hyperspectral Image Classification","volume":"58","author":"Hu","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"6808","DOI":"10.1109\/TGRS.2019.2908756","article-title":"Unsupervised Spatial\u2013Spectral Feature Learning by 3D Convolutional Autoencoder for Hyperspectral Classification","volume":"57","author":"Mei","year":"2019","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_11","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., and Polosukhin, I. (2017). Advances in Neural Information Processing Systems, Curran Associates, Inc."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"847","DOI":"10.1109\/TGRS.2017.2755542","article-title":"Spectral\u2013Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework","volume":"56","author":"Zhong","year":"2018","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Ma, W., Yang, Q., Wu, Y., Zhao, W., and Zhang, X. (2019). Double-Branch Multi-Attention Mechanism Network for Hyperspectral Image Classification. Remote Sens., 11.","DOI":"10.3390\/rs11111307"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"3173","DOI":"10.1109\/TGRS.2018.2794326","article-title":"Hyperspectral Image Classification With Deep Feature Fusion Network","volume":"56","author":"Song","year":"2018","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., and Darrell, T. (2015, January 7\u201312). Fully Convolutional Networks for Semantic Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref_16","first-page":"234","article-title":"U-Net: Convolutional Networks for Biomedical Image Segmentation","volume":"Volume 9351","author":"Navab","year":"2015","journal-title":"Medical Image Computing and Computer-Assisted Intervention\u2014MICCAI 2015"},{"key":"ref_17","first-page":"12077","article-title":"SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers","volume":"Volume 34","author":"Xie","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"492","DOI":"10.1109\/TBDATA.2019.2923243","article-title":"Beyond the Patchwise Classification: Spectral-Spatial Fully Convolutional Networks for Hyperspectral Image Classification","volume":"6","author":"Xu","year":"2020","journal-title":"IEEE Trans. Big Data"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"5612","DOI":"10.1109\/TGRS.2020.2967821","article-title":"FPGA: Fast Patch-Free Global Learning Framework for Fully End-to-End Hyperspectral Image Classification","volume":"58","author":"Zheng","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"107297","DOI":"10.1016\/j.compag.2022.107297","article-title":"HSI-TransUNet: A Transformer Based Semantic Segmentation Model for Crop Mapping from UAV Hyperspectral Imagery","volume":"201","author":"Niu","year":"2022","journal-title":"Comput. Electron. Agric."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"107474","DOI":"10.1016\/j.compag.2022.107474","article-title":"Fine Hyperspectral Classification of Rice Varieties Based on Attention Module 3D-2DCNN","volume":"203","author":"Meng","year":"2022","journal-title":"Comput. Electron. Agric."},{"key":"ref_22","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2021). An Image Is Worth 16 \u00d7 16 Words: Transformers for Image Recognition at Scale. arXiv."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TGRS.2022.3172371","article-title":"SpectralFormer: Rethinking Hyperspectral Image Classification With Transformers","volume":"60","author":"Hong","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.-P., Song, K., Liang, D., Lu, T., Luo, P., and Shao, L. (2021, January 10\u201317). Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"ref_25","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., and Jegou, H. (2021, January 18\u201324). Training Data-Efficient Image Transformers & Distillation through Attention. Proceedings of the 38th International Conference on Machine Learning PMLR, Virtual."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"4729","DOI":"10.1109\/TGRS.2017.2698503","article-title":"Learning and Transferring Deep Joint Spectral\u2013Spatial Features for Hyperspectral Classification","volume":"55","author":"Yang","year":"2017","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"108","DOI":"10.1016\/j.isprsjprs.2017.11.003","article-title":"MugNet: Deep Learning for Hyperspectral Image Classification Using Limited Samples","volume":"145","author":"Pan","year":"2018","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_28","unstructured":"Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv."},{"key":"ref_29","unstructured":"Chu, X., Tian, Z., Zhang, B., Wang, X., Wei, X., Xia, H., and Shen, C. (2021). Conditional Positional Encodings for Vision Transformers. arXiv."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., and Guo, B. (2021, January 10\u201317). Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., and Ahmadi, S.-A. (2016, January 25\u201328). V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA.","DOI":"10.1109\/3DV.2016.79"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Goyal, P., Girshick, R., He, K., and Dollar, P. (2017, January 21\u201326). Focal Loss for Dense Object Detection. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Honolulu, HI, USA.","DOI":"10.1109\/ICCV.2017.324"},{"key":"ref_33","unstructured":"Loshchilov, I., and Hutter, F. (2017). SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"112012","DOI":"10.1016\/j.rse.2020.112012","article-title":"WHU-Hi: UAV-Borne Hyperspectral with High Spatial Resolution (H2) Benchmark Datasets and Classifier for Precise Crop Identification Based on Deep Convolutional Neural Network with CRF","volume":"250","author":"Zhong","year":"2020","journal-title":"Remote Sens. Environ."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/14\/3491\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T20:10:42Z","timestamp":1760127042000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/14\/3491"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,11]]},"references-count":34,"journal-issue":{"issue":"14","published-online":{"date-parts":[[2023,7]]}},"alternative-id":["rs15143491"],"URL":"https:\/\/doi.org\/10.3390\/rs15143491","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,11]]}}}