{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:21:54Z","timestamp":1763644914010,"version":"3.45.0"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T00:00:00Z","timestamp":1761523200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T00:00:00Z","timestamp":1761523200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00371-025-04191-2","type":"journal-article","created":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T01:58:23Z","timestamp":1761530303000},"page":"12889-12903","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Image augmentation by a vision-language foundation model for durian leaf disease recognition"],"prefix":"10.1007","volume":"41","author":[{"given":"Wenjuan","family":"Liu","sequence":"first","affiliation":[]},{"given":"Ahmad Sufril Azlan","family":"Mohamed","sequence":"additional","affiliation":[]},{"given":"Mohd Azam","family":"Osman","sequence":"additional","affiliation":[]},{"given":"Kim Hwa","family":"Kie","sequence":"additional","affiliation":[]},{"given":"Chow Jeng","family":"Wong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"4191_CR1","unstructured":"FAO, Durian Global Trade Overview 2023 (2023)."},{"key":"4191_CR2","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1038\/s41579-023-00900-7","volume":"21","author":"BK Singh","year":"2023","unstructured":"Singh, B.K., Delgado-Baquerizo, M., Egidi, E., Guirado, E., Leach, J.E., Liu, H., Trivedi, P.: Climate change impacts on plant pathogens, food security and paths forward. Nat. Rev. Microbiol. 21, 640\u2013656 (2023). https:\/\/doi.org\/10.1038\/s41579-023-00900-7","journal-title":"Nat. Rev. Microbiol."},{"key":"4191_CR3","doi-asserted-by":"publisher","unstructured":"FAO, Production practices to increase yield, quality and safety of fruits and vegetables (2024). https:\/\/doi.org\/10.4060\/cd0726en.","DOI":"10.4060\/cd0726en"},{"key":"4191_CR4","doi-asserted-by":"publisher","first-page":"108757","DOI":"10.1016\/j.compag.2024.108757","volume":"219","author":"M Ariza-Sent\u00eds","year":"2024","unstructured":"Ariza-Sent\u00eds, M., V\u00e9lez, S., Mart\u00ednez-Pe\u00f1a, R., Baja, H., Valente, J.: Object detection and tracking in precision farming: a systematic review. Comput. Electron. Agric. 219, 108757 (2024). https:\/\/doi.org\/10.1016\/j.compag.2024.108757","journal-title":"Comput. Electron. Agric."},{"key":"4191_CR5","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-025-04022-4","author":"H Lu","year":"2025","unstructured":"Lu, H., Dong, B., Zhu, B., Ma, S., Zhang, Z., Peng, J., Song, K.: A survey on deep learning-based object detection for crop monitoring: pest, yield, weed, and growth applications. Vis. Comput. (2025). https:\/\/doi.org\/10.1007\/s00371-025-04022-4","journal-title":"Vis. Comput."},{"key":"4191_CR6","doi-asserted-by":"publisher","first-page":"92","DOI":"10.1007\/s10462-024-11100-x","volume":"58","author":"A Upadhyay","year":"2025","unstructured":"Upadhyay, A., Chandel, N.S., Singh, K.P., Chakraborty, S.K., Nandede, B.M., Kumar, M., Subeesh, A., Upendar, K., Salem, A., Elbeltagi, A.: Deep learning and computer vision in plant disease detection: a comprehensive review of techniques, models, and trends in precision agriculture. Artif. Intell. Rev. 58, 92 (2025). https:\/\/doi.org\/10.1007\/s10462-024-11100-x","journal-title":"Artif. Intell. Rev."},{"key":"4191_CR7","doi-asserted-by":"publisher","first-page":"4815","DOI":"10.1007\/s00371-024-03692-w","volume":"41","author":"H Zhang","year":"2025","unstructured":"Zhang, H., Ren, G.: Intelligent leaf disease diagnosis: image algorithms using Swin Transformer and federated learning. Vis. Comput. 41, 4815\u20134838 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03692-w","journal-title":"Vis. Comput."},{"key":"4191_CR8","doi-asserted-by":"publisher","first-page":"5639","DOI":"10.1007\/s00371-023-03127-y","volume":"40","author":"BM Joshi","year":"2024","unstructured":"Joshi, B.M., Bhavsar, H.: A nightshade crop leaf disease detection using enhance-nightshade-CNN for ground truth data. Vis. Comput. 40, 5639\u20135658 (2024). https:\/\/doi.org\/10.1007\/s00371-023-03127-y","journal-title":"Vis. Comput."},{"key":"4191_CR9","doi-asserted-by":"publisher","first-page":"107208","DOI":"10.1016\/j.compag.2022.107208","volume":"200","author":"Y Lu","year":"2022","unstructured":"Lu, Y., Chen, D., Olaniyi, E., Huang, Y.: Generative adversarial networks (GANs) for image augmentation in agriculture: a systematic review. Comput. Electron. Agric. 200, 107208 (2022). https:\/\/doi.org\/10.1016\/j.compag.2022.107208","journal-title":"Comput. Electron. Agric."},{"key":"4191_CR10","doi-asserted-by":"publisher","unstructured":"Boondach, M., Chaisiriprasert, P.: Improving Durian Leaf Disease Detection Using LAB Color Space and CLAHE Technique with YOLOv8 Integration. In: 2024 8th International Conference on Information Technology (InCIT) (IEEE, 2024), pp. 162\u2013167. https:\/\/doi.org\/10.1109\/InCIT63192.2024.10810589.","DOI":"10.1109\/InCIT63192.2024.10810589"},{"key":"4191_CR11","doi-asserted-by":"publisher","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks. In: Proceedings of the 28th International Conference on Neural Information Processing Systems, 63: 139\u2013144 (2014). https:\/\/doi.org\/10.1145\/3422622.","DOI":"10.1145\/3422622"},{"key":"4191_CR12","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: International Conference on Machine Learning, pp. 214\u2013223 (2017)."},{"issue":"2024","key":"4191_CR13","doi-asserted-by":"publisher","first-page":"3971","DOI":"10.1007\/s00371-024-03404-4","volume":"40","author":"W Zhao","year":"2024","unstructured":"Zhao, W., Zhu, J., Li, P., Huang, J., Tang, J.: Attention mechanism-based generative adversarial networks for image cartoonization. Vis. Comput. 40(2024), 3971\u20133984 (2024). https:\/\/doi.org\/10.1007\/s00371-024-03404-4","journal-title":"Vis. Comput."},{"key":"4191_CR14","doi-asserted-by":"publisher","first-page":"2299","DOI":"10.1007\/s00371-024-03534-9","volume":"41","author":"R Elanwar","year":"2025","unstructured":"Elanwar, R., Betke, M.: Generative adversarial networks for handwriting image generation: a review. Vis. Comput. 41, 2299\u20132322 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03534-9","journal-title":"Vis. Comput."},{"key":"4191_CR15","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising Diffusion Probabilistic Models (2020)."},{"key":"4191_CR16","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat gans on image synthesis. In: Advances in neural information processing systems, pp. 8780\u20138794 (2021)."},{"key":"4191_CR17","unstructured":"Saharia, C., Chan, W., Saxena, S., et al.: Photorealistic text-to-image diffusion models with deep language understanding. In: Advances in neural information processing systems, pp. 36479\u201336494 (2022)."},{"key":"4191_CR18","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-Resolution Image Synthesis with Latent Diffusion Models (2021).","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"4191_CR19","doi-asserted-by":"publisher","first-page":"4481","DOI":"10.1007\/s00371-024-03673-z","volume":"41","author":"H Wu","year":"2025","unstructured":"Wu, H., Li, B., Tian, L., Dong, C.: DDFA: a displacement and diffusion-based feature augmentation method for imbalanced image recognition. Vis. Comput. 41, 4481\u20134495 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03673-z","journal-title":"Vis. Comput."},{"key":"4191_CR20","unstructured":"OpenAI, DALL\u00b7E 3. https:\/\/openai.com\/index\/dall-e-3\/, (2023)."},{"key":"4191_CR21","unstructured":"Stability AI, Introducing SD3. https:\/\/www.diffus.me\/, pp. 6\u201312 (2024)."},{"key":"4191_CR22","unstructured":"BlackForest Labs, Introducing Flux1. https:\/\/blackforestlabs.ai\/, (2024)."},{"key":"4191_CR23","doi-asserted-by":"publisher","first-page":"2953","DOI":"10.1007\/s00371-024-03579-w","volume":"41","author":"C Liu","year":"2025","unstructured":"Liu, C., Jin, Y., Guan, Z., Li, T., Qin, Y., Qian, B., Jiang, Z., Wu, Y., Wang, X., Zheng, Y.F., Zeng, D.: Visual\u2013language foundation models in medicine. Vis. Comput. 41, 2953\u20132972 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03579-w","journal-title":"Vis. Comput."},{"key":"4191_CR24","doi-asserted-by":"publisher","first-page":"109032","DOI":"10.1016\/j.compag.2024.109032","volume":"222","author":"J Li","year":"2024","unstructured":"Li, J., Xu, M., Xiang, L., Chen, D., Zhuang, W., Yin, X., Li, Z.: Foundation models in smart agriculture: basics, opportunities, and challenges. Comput. Electron. Agric. 222, 109032 (2024). https:\/\/doi.org\/10.1016\/j.compag.2024.109032","journal-title":"Comput. Electron. Agric."},{"key":"4191_CR25","unstructured":"Hu, E.J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W.: LoRA: Low-Rank Adaptation of Large Language Models (2021)."},{"key":"4191_CR26","unstructured":"Gal, R., Alaluf, Y., Atzmon, Y., Patashnik, O., Bermano, A.H., Chechik, G., Cohen-Or, D.: An image is worth one word: personalizing text-to-image generation using textual inversion (2022)."},{"key":"4191_CR27","doi-asserted-by":"publisher","first-page":"2337","DOI":"10.1007\/s11263-022-01653-1","volume":"130","author":"K Zhou","year":"2022","unstructured":"Zhou, K., Yang, J., Loy, C.C., Liu, Z.: Learning to prompt for vision-language models. Int. J. Comput. Vis. 130, 2337\u20132348 (2022). https:\/\/doi.org\/10.1007\/s11263-022-01653-1","journal-title":"Int. J. Comput. Vis."},{"key":"4191_CR28","unstructured":"Lipman, Y., Chen, R.T.Q., Ben-Hamu, H., Nickel, M., Le, M.: Flow Matching for Generative Modeling (2022)."},{"key":"4191_CR29","unstructured":"Liu, X., Gong, C., Liu, Q.: Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow (2022)."},{"key":"4191_CR30","unstructured":"Liu, X., Zhang, X., Ma, J., Peng, J., Liu, Q.: InstaFlow: One Step is Enough for High-Quality Diffusion-Based Text-to-Image Generation (2023)."},{"key":"4191_CR31","unstructured":"Wang, J., Pu, J., Qi, Z., Guo, J., Ma, Y., Huang, N., Chen, Y., Li, X., Shan, Y.: Taming Rectified Flow for Inversion and Editing (2025)."},{"key":"4191_CR32","unstructured":"Jin, Y., Sun, Z., Li, N., Xu, K., Xu, K., Jiang, H., Zhuang, N., Huang, Q., Song, Y., Mu, Y., Lin, Z.: Pyramidal Flow Matching for Efficient Video Generative Modeling (2025)."},{"key":"4191_CR33","unstructured":"Cao, Y., Song, Z., Yang, C.: Video Latent Flow Matching: Optimal Polynomial Projections for Video Interpolation and Extrapolation (2025)."},{"key":"4191_CR34","doi-asserted-by":"publisher","unstructured":"Schusterbauer, J., Gui, M., Ma, P., Stracke, N., Baumann, S.A., Hu, V.T., Ommer, B.: FMBoost: Boosting Latent Diffusion with\u00a0Flow Matching, pp. 338\u2013355 (2025). https:\/\/doi.org\/10.1007\/978-3-031-73030-6_19.","DOI":"10.1007\/978-3-031-73030-6_19"},{"key":"4191_CR35","unstructured":"OpenAI, Gpt-4 technical report. https:\/\/cdn.openai.com\/papers\/gpt-4.pdf, (2023)."},{"key":"4191_CR36","unstructured":"Esser, P., Kulal, S., Blattmann, A., Entezari, R., M\u00fcller, J., Saini, H., Levi, Y., Lorenz, D., Sauer, A., Boesel, F., Podell, D., Dockhorn, T., English, Z., Lacey, K., Goodwin, A., Marek, Y., Rombach, R: Scaling Rectified Flow Transformers for High-Resolution Image Synthesis (2024)."},{"key":"4191_CR37","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale (2020)."},{"key":"4191_CR38","doi-asserted-by":"crossref","unstructured":"Ben Zaken, E., Ravfogel, S., Goldberg, Y.: BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models (2022).","DOI":"10.18653\/v1\/2022.acl-short.1"},{"key":"4191_CR39","unstructured":"Zhang, R., Han, J., Liu, C., Gao, P., Zhou, A., Hu, X., Yan, S., Lu, P., Li, H., Qiao, Y.: LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention (2024)."},{"key":"4191_CR40","unstructured":"Houlsby, N., Giurgiu, A., Jastrzebski, S., Morrone, B., De Laroussilhe, Q., Gesmundo, A., Attariyan, M., Gelly, S.: Parameter-efficient transfer learning for NLP. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proceedings of the 36th International Conference on Machine Learning, pp. 2790\u20132799 (PMLR, 2019)."},{"key":"4191_CR41","unstructured":"Yeh, S.-Y., Hsieh, Y.-G., Gao, Z., Yang, B.B.W., Oh, G., Gong, Y.: Navigating Text-To-Image Customization: From LyCORIS Fine-Tuning to Model Evaluation (2023)."},{"key":"4191_CR42","unstructured":"Black Forest Labs, Black Forest Labs. FLUX.1: Parameter-Efficient LoRA Fine-tuning Documentation. https:\/\/huggingface.co\/black-forest-labs\/FLUX.1-dev#parameter-efficient-lora-fine-tuning, (2024)."},{"key":"4191_CR43","unstructured":"van den Oord, A.., Li, Y., Vinyals, O.: Representation Learning with Contrastive Predictive Coding (2019)."},{"key":"4191_CR44","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A Simple Framework for Contrastive Learning of Visual Representations. In: III, H.D., Singh, A. (eds.) Proceedings of the 37th International Conference on Machine Learning, pp. 1597\u20131607. (PMLR, 2020)"},{"key":"4191_CR45","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., Sutskever, I.: Learning Transferable Visual Models From Natural Language Supervision (2021)."},{"key":"4191_CR46","unstructured":"Jia, C., Yang, Y., Xia, Y., Chen, Y.-T., Parekh, Z., Pham, H., Le, Q.V., Sung, Y., Li, Z.,. Duerig, T.: Scaling up visual and vision-language representation learning with noisy text supervision (2021)."},{"key":"4191_CR47","unstructured":"Black Forest Labs, FLUX.1-dev Model. https:\/\/huggingface.co\/black-forest-labs\/FLUX.1-dev, (2024)."},{"key":"4191_CR48","unstructured":"Kingma, D.P., Welling, M.: Auto-Encoding Variational Bayes (2013)."},{"key":"4191_CR49","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee. K., Narang, S., Matena, M., Zhou, Y., Li, W., Liu, P.J.: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 1\u201321 (2020)"},{"key":"4191_CR50","unstructured":"Roboflow, Durian Diseases. https:\/\/universe.roboflow.com\/machine-vision-leaf-disease\/durian-disease-llewn-cmng1, (2024)."},{"key":"4191_CR51","unstructured":"Bi\u0144kowski, M., Sutherland, D.J., Arbel, M., Gretton, A.: Demystifying MMD GANs. (2018)."},{"key":"4191_CR52","first-page":"723","volume":"13","author":"A Gretton","year":"2012","unstructured":"Gretton, A., Borgwardt, K.M., Rasch, M.J.: A kernel two-sample test. J. Mach. Learn. Res. 13, 723\u2013773 (2012)","journal-title":"J. Mach. Learn. Res."},{"key":"4191_CR53","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, (2017)."},{"key":"4191_CR54","doi-asserted-by":"crossref","unstructured":"Jayasumana, S., Ramalingam, S., Veit, A.: Rethinking FID Towards a Better Evaluation Metric for Image Generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 9307\u20139315 (2024)","DOI":"10.1109\/CVPR52733.2024.00889"},{"key":"4191_CR55","unstructured":"Ravuri, S., Vinyals, O.: Classification Accuracy Score for Conditional Generative Models (2019)."},{"key":"4191_CR56","unstructured":"Jocher, G., Chaurasia, A., Qiu, J.: Ultralytics YOLOv8. https:\/\/github.com\/ultralytics\/ultralytics (2023)."},{"key":"4191_CR57","unstructured":"Arora, S., Hu, W., Pravesh, K., Kothari: An analysis of the t-SNE algorithm for data visualization. In: Proceedings of the 31st conference on learning theory, 1455\u20131462. PMLR, (2018)."},{"key":"4191_CR58","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., et al.: Grad-cam: Visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision (2017).","DOI":"10.1109\/ICCV.2017.74"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04191-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04191-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04191-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:16:32Z","timestamp":1763644592000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04191-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"references-count":58,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4191"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04191-2","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"14 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}