{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T19:27:19Z","timestamp":1769196439378,"version":"3.49.0"},"reference-count":47,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Engineering Applications of Artificial Intelligence"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1016\/j.engappai.2025.113588","type":"journal-article","created":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T19:19:33Z","timestamp":1766863173000},"page":"113588","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"PB","title":["Triplex attention enabled diffusion generative adversarial network for photorealistic image generation from text and images"],"prefix":"10.1016","volume":"166","author":[{"given":"Sonal S.","family":"Fatangare","sequence":"first","affiliation":[]},{"given":"Premanand P.","family":"Ghadekar","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.engappai.2025.113588_bib1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s40537-021-00444-8","article-title":"Review of deep learning: concepts, CNN architectures, challenges, applications, future directions","volume":"8","author":"Alzubaidi","year":"2021","journal-title":"J. Big Data"},{"key":"10.1016\/j.engappai.2025.113588_bib2","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.108998","article-title":"A full data augmentation pipeline for small object detection based on generative adversarial networks","volume":"133","author":"Bosquet","year":"2023","journal-title":"Pattern Recogn."},{"key":"10.1016\/j.engappai.2025.113588_bib3","doi-asserted-by":"crossref","first-page":"22051","DOI":"10.1109\/ACCESS.2023.3251988","article-title":"A novel data augmentation method for improved visual crack detection using generative adversarial networks","volume":"11","author":"Branikas","year":"2023","journal-title":"IEEE Access"},{"key":"10.1016\/j.engappai.2025.113588_bib4","series-title":"A Comprehensive Survey of AI-generated Content (AIGC): a History of Generative AI from GAN to ChatGPT","author":"Cao","year":"2023"},{"key":"10.1016\/j.engappai.2025.113588_bib5","doi-asserted-by":"crossref","DOI":"10.1109\/ACCESS.2024.3491017","article-title":"Robustness of generative adversarial CLIPs against single-character adversarial attacks in text-to-image generation","author":"Chanakya","year":"2024","journal-title":"IEEE Access"},{"key":"10.1016\/j.engappai.2025.113588_bib6","doi-asserted-by":"crossref","first-page":"4555","DOI":"10.1007\/s11263-025-02371-0","article-title":"Attribute-centric compositional text-to-image generation","volume":"133","author":"Cong","year":"2025","journal-title":"Int. J. Comput. Vis."},{"issue":"2","key":"10.1016\/j.engappai.2025.113588_bib8","doi-asserted-by":"crossref","first-page":"30","DOI":"10.3390\/computers8020030","article-title":"A study of image upsampling and downsampling filters","volume":"8","author":"Dumitrescu","year":"2019","journal-title":"Computers"},{"issue":"5","key":"10.1016\/j.engappai.2025.113588_bib9","doi-asserted-by":"crossref","first-page":"4609","DOI":"10.1007\/s11063-022-10777-x","article-title":"Image generation: a review","volume":"54","author":"Elasri","year":"2022","journal-title":"Neural Process. Lett."},{"key":"10.1016\/j.engappai.2025.113588_bib10","doi-asserted-by":"crossref","first-page":"2112","DOI":"10.1109\/TMM.2021.3075997","article-title":"Modality disentangled discriminator for text-to-image synthesis","volume":"24","author":"Feng","year":"2021","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.engappai.2025.113588_bib12","series-title":"Evaluating the Robustness of text-to-image Diffusion Models Against real-world Attacks","author":"Gao","year":"2023"},{"key":"10.1016\/j.engappai.2025.113588_bib13","series-title":"Proc. IEEE\/CVF Int. Conf. Comput. Vis. (ICCV)","first-page":"8059","article-title":"Boosting few-shot visual learning with self-supervision","author":"Gidaris","year":"2019"},{"issue":"11","key":"10.1016\/j.engappai.2025.113588_bib14","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","article-title":"Generative adversarial networks","volume":"63","author":"Goodfellow","year":"2020","journal-title":"Commun. ACM"},{"key":"10.1016\/j.engappai.2025.113588_bib15","doi-asserted-by":"crossref","first-page":"9572","DOI":"10.1109\/ACCESS.2023.3342866","article-title":"GACnet-Text-to-Image synthesis with generative models using attention mechanisms with contrastive learning","volume":"12","author":"Habib","year":"2023","journal-title":"IEEE Access"},{"issue":"3","key":"10.1016\/j.engappai.2025.113588_bib16","doi-asserted-by":"crossref","first-page":"1552","DOI":"10.1109\/TPAMI.2020.3021209","article-title":"Semantic object accuracy for generative text-to-image synthesis","volume":"44","author":"Hinz","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.engappai.2025.113588_bib17","series-title":"Towards Text Contextual Understanding: Text Feature Fusion GAN for text-to-image Generation","author":"Jiang","year":"2024"},{"key":"10.1016\/j.engappai.2025.113588_bib18","doi-asserted-by":"crossref","first-page":"88152","DOI":"10.1109\/ACCESS.2023.3305994","article-title":"Generating chest X-Ray progression of pneumonia using conditional cycle generative adversarial networks","volume":"11","author":"Jin","year":"2023","journal-title":"IEEE Access"},{"key":"10.1016\/j.engappai.2025.113588_bib19","doi-asserted-by":"crossref","first-page":"1250","DOI":"10.1109\/ACCESS.2020.3015656","article-title":"A realistic image generation of face from text description using the fully trained generative adversarial networks","volume":"9","author":"Khan","year":"2020","journal-title":"IEEE Access"},{"issue":"8","key":"10.1016\/j.engappai.2025.113588_bib20","doi-asserted-by":"crossref","first-page":"5098","DOI":"10.3390\/app13085098","article-title":"TextControlGAN: text-to-image synthesis with controllable generative adversarial networks","volume":"13","author":"Ku","year":"2023","journal-title":"Appl. Sci."},{"key":"10.1016\/j.engappai.2025.113588_bib21","first-page":"38685","article-title":"Conditional score guidance for text-driven image-to-image translation","volume":"36","author":"Lee","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"2","key":"10.1016\/j.engappai.2025.113588_bib22","doi-asserted-by":"crossref","first-page":"697","DOI":"10.2991\/ijcis.d.190617.001","article-title":"3D model generation and reconstruction using a conditional generative adversarial network","volume":"12","author":"Li","year":"2019","journal-title":"Int. J. Comput. Intell. Syst."},{"issue":"8","key":"10.1016\/j.engappai.2025.113588_bib23","first-page":"1","article-title":"Mobile user traffic generation via multi-scale hierarchical GAN","volume":"18","author":"Li","year":"2024","journal-title":"ACM Trans. Knowl. Discov. Data"},{"key":"10.1016\/j.engappai.2025.113588_bib24","doi-asserted-by":"crossref","first-page":"977","DOI":"10.1016\/j.procs.2018.04.239","article-title":"Research on convolutional neural network based on improved ReLU piecewise activation function","volume":"131","author":"Lin","year":"2018","journal-title":"Procedia Comput. Sci."},{"key":"10.1016\/j.engappai.2025.113588_bib25","article-title":"Unsupervised image-to-image translation networks","volume":"30","author":"Liu","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"5","key":"10.1016\/j.engappai.2025.113588_bib26","doi-asserted-by":"crossref","first-page":"839","DOI":"10.1109\/JPROC.2021.3049196","article-title":"Generative adversarial networks for image and video synthesis: algorithms and applications","volume":"109","author":"Liu","year":"2021","journal-title":"Proc. IEEE"},{"issue":"15","key":"10.1016\/j.engappai.2025.113588_bib27","doi-asserted-by":"crossref","first-page":"6831","DOI":"10.3390\/app14156831","article-title":"Synthetic medical imaging generation with generative adversarial networks for plain radiographs","volume":"14","author":"McNulty","year":"2024","journal-title":"Appl. Sci."},{"issue":"1","key":"10.1016\/j.engappai.2025.113588_bib28","first-page":"84","article-title":"Generative adversarial networks for the synthesis of chest X-ray images","volume":"31","author":"Ng","year":"2023","journal-title":"Eng. Proc."},{"key":"10.1016\/j.engappai.2025.113588_bib29","doi-asserted-by":"crossref","first-page":"9091","DOI":"10.1007\/s00371-025-03915-8","article-title":"Quantum generative adversarial network for image generation","volume":"41","author":"Pajuhanfard","year":"2025","journal-title":"Vis. Comput."},{"key":"10.1016\/j.engappai.2025.113588_bib30","doi-asserted-by":"crossref","DOI":"10.1016\/j.artmed.2023.102556","article-title":"Progressive growing of generative adversarial networks for improving data augmentation and skin cancer diagnosis","volume":"141","author":"P\u00e9rez","year":"2023","journal-title":"Artif. Intell. Med."},{"key":"10.1016\/j.engappai.2025.113588_bib31","series-title":"Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recogn","first-page":"7932","article-title":"Grounded text-to-image synthesis with attention refocusing","author":"Phung","year":"2024"},{"key":"10.1016\/j.engappai.2025.113588_bib32","series-title":"Hierarchical text-conditional Image Generation with CLIP Latents","author":"Ramesh","year":"2022"},{"issue":"1","key":"10.1016\/j.engappai.2025.113588_bib33","first-page":"16","article-title":"Text-to-Image generation using deep learning","volume":"20","author":"Ramzan","year":"2022","journal-title":"Eng. Proc."},{"key":"10.1016\/j.engappai.2025.113588_bib34","doi-asserted-by":"crossref","DOI":"10.1109\/TVCG.2024.3456213","article-title":"Expressive 3D facial animation generation based on local-to-global latent diffusion","author":"Song","year":"2024","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"10.1016\/j.engappai.2025.113588_bib35","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2025.111447","article-title":"AttriDiffuser: adversarially enhanced diffusion model for text-to-facial attribute image synthesis","volume":"163","author":"Song","year":"2025","journal-title":"Pattern Recogn."},{"issue":"2","key":"10.1016\/j.engappai.2025.113588_bib36","first-page":"832","article-title":"Cross-modal semantic matching generative adversarial networks for text-to-image synthesis","volume":"24","author":"Tan","year":"2021","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.engappai.2025.113588_bib37","doi-asserted-by":"crossref","first-page":"39508","DOI":"10.1109\/ACCESS.2023.3268869","article-title":"Enhanced text-to-image synthesis with self-supervision","volume":"11","author":"Tan","year":"2023","journal-title":"IEEE Access"},{"key":"10.1016\/j.engappai.2025.113588_bib38","series-title":"Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recogn","first-page":"16515","article-title":"DF-GAN: a simple and effective baseline for text-to-image synthesis","author":"Tao","year":"2022"},{"issue":"4","key":"10.1016\/j.engappai.2025.113588_bib39","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3658157","article-title":"Training-free consistent text-to-image generation","volume":"43","author":"Tewel","year":"2024","journal-title":"ACM Trans. Graph."},{"key":"10.1016\/j.engappai.2025.113588_bib40","series-title":"Diffusion-GAN: Training GANs with Diffusion","author":"Wang","year":"2022"},{"key":"10.1016\/j.engappai.2025.113588_bib41","series-title":"Personalized Image Generation with Deep Generative Models: a Decade Survey","author":"Wei","year":"2025"},{"issue":"1","key":"10.1016\/j.engappai.2025.113588_bib42","doi-asserted-by":"crossref","DOI":"10.1080\/08839514.2021.1988441","article-title":"Highlight removal from a single grayscale image using attentive GAN","volume":"36","author":"Xu","year":"2022","journal-title":"Appl. Artif. Intell."},{"key":"10.1016\/j.engappai.2025.113588_bib43","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"1316","article-title":"AttnGAN: fine-grained text to image generation with attentional generative adversarial networks","author":"Xu","year":"2018"},{"key":"10.1016\/j.engappai.2025.113588_bib44","series-title":"On the multi-modal Vulnerability of Diffusion Models","author":"Yang","year":"2024"},{"issue":"5","key":"10.1016\/j.engappai.2025.113588_bib45","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s40747-025-01832-w","article-title":"A study of enhanced visual perception of marine biology images based on diffusion-GAN","volume":"11","author":"Yao","year":"2025","journal-title":"Complex & Intell. Syst."},{"issue":"1","key":"10.1016\/j.engappai.2025.113588_bib46","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1007\/s11063-024-11503-5","article-title":"RII-GAN: multi-scaled aligning-based reversed image interaction network for text-to-image synthesis","volume":"56","author":"Yuan","year":"2024","journal-title":"Neural Process. Lett."},{"issue":"8","key":"10.1016\/j.engappai.2025.113588_bib47","doi-asserted-by":"crossref","first-page":"1947","DOI":"10.1109\/TPAMI.2018.2856256","article-title":"StackGAN++: realistic image synthesis with stacked generative adversarial networks","volume":"41","author":"Zhang","year":"2018","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"4","key":"10.1016\/j.engappai.2025.113588_bib48","doi-asserted-by":"crossref","first-page":"1113","DOI":"10.1109\/TETCI.2023.3243920","article-title":"Multi-scale attention generative adversarial network for medical image enhancement","volume":"7","author":"Zhong","year":"2023","journal-title":"IEEE Trans. Emerg. Top. Comput. Intell."},{"issue":"6","key":"10.1016\/j.engappai.2025.113588_bib49","doi-asserted-by":"crossref","first-page":"1205","DOI":"10.1007\/s41095-024-0422-4","article-title":"Multi3D: 3D-aware multimodal image synthesis","volume":"10","author":"Zhou","year":"2024","journal-title":"Comput. Vis. Media."}],"container-title":["Engineering Applications of Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625036206?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625036206?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T16:19:21Z","timestamp":1769185161000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0952197625036206"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":47,"alternative-id":["S0952197625036206"],"URL":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113588","relation":{},"ISSN":["0952-1976"],"issn-type":[{"value":"0952-1976","type":"print"}],"subject":[],"published":{"date-parts":[[2026,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Triplex attention enabled diffusion generative adversarial network for photorealistic image generation from text and images","name":"articletitle","label":"Article Title"},{"value":"Engineering Applications of Artificial Intelligence","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113588","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"113588"}}