{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,29]],"date-time":"2025-09-29T11:41:31Z","timestamp":1759146091841,"version":"3.38.0"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T00:00:00Z","timestamp":1726272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T00:00:00Z","timestamp":1726272000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62366018"],"award-info":[{"award-number":["62366018"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004479","name":"Natural Science Foundation of Jiangxi Province","doi-asserted-by":"publisher","award":["20212BAB212005, 20224BAB202015, 20224BAB202020"],"award-info":[{"award-number":["20212BAB212005, 20224BAB202015, 20224BAB202020"]}],"id":[{"id":"10.13039\/501100004479","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100017355","name":"Graduate Innovative Special Fund Projects of Jiangxi Province","doi-asserted-by":"publisher","award":["YC2023-S365"],"award-info":[{"award-number":["YC2023-S365"]}],"id":[{"id":"10.13039\/100017355","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s00371-024-03617-7","type":"journal-article","created":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T15:02:03Z","timestamp":1726326123000},"page":"3519-3535","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Text-guided floral image generation based on lightweight deep attention feature fusion GAN"],"prefix":"10.1007","volume":"41","author":[{"given":"Wenji","family":"Yang","sequence":"first","affiliation":[]},{"given":"Hang","family":"An","sequence":"additional","affiliation":[]},{"given":"Wenchao","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Xinxin","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Liping","family":"Xie","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,14]]},"reference":[{"issue":"4","key":"3617_CR1","doi-asserted-by":"publisher","first-page":"1099","DOI":"10.1007\/s11252-023-01353-9","volume":"26","author":"JM Schmack","year":"2023","unstructured":"Schmack, J.M., Egerer, M.: Floral richness and seasonality influences bee and non-bee flower interactions in urban community gardens. Urban Ecosyst. 26(4), 1099\u20131112 (2023)","journal-title":"Urban Ecosyst."},{"key":"3617_CR2","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1016\/j.jiec.2023.02.013","volume":"122","author":"F Kaya","year":"2023","unstructured":"Kaya, F., Solmaz, R., Ge\u00e7ibesler, \u0130H.: The use of methanol extract of rheum ribes (i\u015fg\u0131n) flower as a natural and promising corrosion inhibitor for mild steel protection in 1 m hcl solution. J. Ind. Eng. Chem. 122, 102\u2013117 (2023)","journal-title":"J. Ind. Eng. Chem."},{"issue":"4","key":"3617_CR3","doi-asserted-by":"publisher","first-page":"2136","DOI":"10.1007\/s00344-022-10688-0","volume":"42","author":"A Chandel","year":"2023","unstructured":"Chandel, A., Thakur, M., Singh, G., Dogra, R., Bajad, A., Soni, V., Bhargava, B.: Flower regulation in floriculture: an agronomic concept and commercial use. J. Plant Growth Regul. 42(4), 2136\u20132161 (2023)","journal-title":"J. Plant Growth Regul."},{"issue":"2","key":"3617_CR4","doi-asserted-by":"publisher","first-page":"1489","DOI":"10.1109\/TPAMI.2022.3164083","volume":"45","author":"Y Li","year":"2022","unstructured":"Li, Y., Yao, T., Pan, Y., Mei, T.: Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 45(2), 1489\u20131500 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3617_CR5","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat gans on image synthesis. Adv. Neural. Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3617_CR6","doi-asserted-by":"crossref","unstructured":"Tao, M., Bao, B.-K., Tang, H., Xu, C.: Galip: Generative adversarial clips for text-to-image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14214\u201314223 (2023)","DOI":"10.1109\/CVPR52729.2023.01366"},{"key":"3617_CR7","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., Aberman, K.: Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22500\u201322510 (2023)","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"3617_CR8","doi-asserted-by":"crossref","unstructured":"Touvron, H., Cord, M., J\u00e9gou, H.: Deit iii: Revenge of the vit. In: European Conference on Computer Vision, pp. 516\u2013533 (2022). Springer","DOI":"10.1007\/978-3-031-20053-3_30"},{"key":"3617_CR9","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"3617_CR10","doi-asserted-by":"crossref","unstructured":"Gu, S., Chen, D., Bao, J., Wen, F., Zhang, B., Chen, D., Yuan, L., Guo, B.: Vector quantized diffusion model for text-to-image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10696\u201310706 (2022)","DOI":"10.1109\/CVPR52688.2022.01043"},{"key":"3617_CR11","unstructured":"Ramesh, A., Pavlov, M., Goh, G., Gray, S., Voss, C., Radford, A., Chen, M., Sutskever, I.: Zero-shot text-to-image generation. In: International Conference on Machine Learning, pp. 8821\u20138831 (2021). Pmlr"},{"key":"3617_CR12","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763 (2021). PMLR"},{"key":"3617_CR13","unstructured":"Nichol, A., Dhariwal, P., Ramesh, A., Shyam, P., Mishkin, P., McGrew, B., Sutskever, I., Chen, M.: Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741 (2021)"},{"issue":"8","key":"3617_CR14","doi-asserted-by":"publisher","first-page":"5187","DOI":"10.1109\/TCSVT.2021.3136857","volume":"32","author":"J Cheng","year":"2021","unstructured":"Cheng, J., Wu, F., Tian, Y., Wang, L., Tao, D.: Rifegan2: Rich feature generation for text-to-image synthesis from constrained prior knowledge. IEEE Trans. Circuits Syst. Video Technol. 32(8), 5187\u20135200 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3617_CR15","unstructured":"Li, D., Li, J., Hoi, S.: Blip-diffusion: Pre-trained subject representation for controllable text-to-image generation and editing. Adv. Neural Inf. Process. Syst. 36 (2024)"},{"key":"3617_CR16","doi-asserted-by":"crossref","unstructured":"Chen, J., Yu, J., Ge, C., Yao, L., Xie, E., Wu, Y., Wang, Z., Kwok, J., Luo, P., Lu, H., et al.: Pixart-$$\\alpha $$: Fast training of diffusion transformer for photorealistic text-to-image synthesis. arXiv preprint arXiv:2310.00426 (2023)","DOI":"10.1007\/978-3-031-73411-3_5"},{"key":"3617_CR17","unstructured":"Bao, F., Nie, S., Xue, K., Li, C., Pu, S., Wang, Y., Yue, G., Cao, Y., Su, H., Zhu, J.: One transformer fits all distributions in multi-modal diffusion at scale. In: International Conference on Machine Learning, pp. 1692\u20131717 (2023). PMLR"},{"key":"3617_CR18","doi-asserted-by":"crossref","unstructured":"Xie, J., Li, Y., Huang, Y., Liu, H., Zhang, W., Zheng, Y., Shou, M.Z.: Boxdiff: Text-to-image synthesis with training-free box-constrained diffusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7452\u20137461 (2023)","DOI":"10.1109\/ICCV51070.2023.00685"},{"key":"3617_CR19","doi-asserted-by":"crossref","unstructured":"Xu, X., Guo, J., Wang, Z., Huang, G., Essa, I., Shi, H.: Prompt-free diffusion: Taking\" text\" out of text-to-image diffusion models. arXiv preprint arXiv:2305.16223 (2023)","DOI":"10.1109\/CVPR52733.2024.00829"},{"issue":"8","key":"3617_CR20","doi-asserted-by":"publisher","first-page":"4499","DOI":"10.1109\/TNNLS.2021.3116209","volume":"34","author":"Z Xie","year":"2021","unstructured":"Xie, Z., Zhang, W., Sheng, B., Li, P., Chen, C.P.: Bagfn: broad attentive graph fusion network for high-order feature interactions. IEEE Trans. Neural Netw. Learn. Syst. 34(8), 4499\u20134513 (2021)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"3617_CR21","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2021","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: Eapt: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"3617_CR22","unstructured":"Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., Lee, H.: Generative adversarial text to image synthesis. In: International Conference on Machine Learning, pp. 1060\u20131069 (2016). PMLR"},{"key":"3617_CR23","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., Metaxas, D.N.: Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5907\u20135915 (2017)","DOI":"10.1109\/ICCV.2017.629"},{"issue":"8","key":"3617_CR24","doi-asserted-by":"publisher","first-page":"1947","DOI":"10.1109\/TPAMI.2018.2856256","volume":"41","author":"H Zhang","year":"2018","unstructured":"Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., Metaxas, D.N.: Stackgan++: Realistic image synthesis with stacked generative adversarial networks. IEEE Trans. Pattern Anal. Mach. Intell. 41(8), 1947\u20131962 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3617_CR25","doi-asserted-by":"crossref","unstructured":"Xu, T., Zhang, P., Huang, Q., Zhang, H., Gan, Z., Huang, X., He, X.: Attngan: Fine-grained text to image generation with attentional generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1316\u20131324 (2018)","DOI":"10.1109\/CVPR.2018.00143"},{"key":"3617_CR26","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Xie, Y., Yang, L.: Photographic text-to-image synthesis with a hierarchically-nested adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6199\u20136208 (2018)","DOI":"10.1109\/CVPR.2018.00649"},{"key":"3617_CR27","doi-asserted-by":"crossref","unstructured":"Tao, M., Tang, H., Wu, F., Jing, X.-Y., Bao, B.-K., Xu, C.: Df-gan: A simple and effective baseline for text-to-image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16515\u201316525 (2022)","DOI":"10.1109\/CVPR52688.2022.01602"},{"key":"3617_CR28","doi-asserted-by":"crossref","unstructured":"Ye, S., Wang, H., Tan, M., Liu, F.: Recurrent affine transformation for text-to-image synthesis. IEEE Trans. Multimedia (2023)","DOI":"10.1109\/TMM.2023.3266607"},{"key":"3617_CR29","doi-asserted-by":"crossref","unstructured":"Liao, W., Hu, K., Yang, M.Y., Rosenhahn, B.: Text to image generation with semantic-spatial aware gan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18187\u201318196 (2022)","DOI":"10.1109\/CVPR52688.2022.01765"},{"key":"3617_CR30","unstructured":"Liu, B., Zhu, Y., Song, K., Elgammal, A.: Towards faster and stabilized gan training for high-fidelity few-shot image synthesis. In: International Conference on Learning Representations (2020)"},{"key":"3617_CR31","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"3617_CR32","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training gans. Adv. Neural Inf. Process. Syst. 29 (2016)"},{"key":"3617_CR33","unstructured":"Reed, S.E., Akata, Z., Mohan, S., Tenka, S., Schiele, B., Lee, H.: Learning what and where to draw. Adv. Neural Inf. Process. Syst. 29 (2016)"},{"key":"3617_CR34","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03617-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03617-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03617-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,10]],"date-time":"2025-03-10T09:10:53Z","timestamp":1741597853000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03617-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,14]]},"references-count":34,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["3617"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03617-7","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2024,9,14]]},"assertion":[{"value":"22 August 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 September 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or nonfinancial interest in the subject matter or materials discussed in this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}