{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T17:25:59Z","timestamp":1771003559353,"version":"3.50.1"},"reference-count":40,"publisher":"SAGE Publications","issue":"2","license":[{"start":{"date-parts":[[2024,12,10]],"date-time":"2024-12-10T00:00:00Z","timestamp":1733788800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Journal of Computational Methods in Sciences and Engineering"],"published-print":{"date-parts":[[2025,3]]},"abstract":"<jats:p>In the process of generating personalized poster content, there are often problems such as inconsistency, blurring, and distortion between the generated image and the original image content, exaggerated image styles that do not meet the personalized needs of users. In response to these issues, this article aims to use Pix2Pix Conditional Generative Adversarial Network (Pix2Pix-CGAN) to solve problems such as image blur, distortion, and content inconsistency in personalized poster generation, in order to improve the quality of generated images and meet users\u2019 personalized needs. This article combined L1 loss and adversarial loss optimization models, proposed a style transfer mechanism for personalized poster generation models based on Pix2Pix, and proposed corresponding detail enhancement strategies. The U-Net architecture was used to improve the generator structure, and PatchGAN was used to ensure local detail clarity and image quality. The experimental results showed that the color consistency (CC) of the personalized poster generation model based on Pix2Pix was 0.97, the Style Preservation Index (SPI) was 0.93, the PSNR was 35\u00a0dB, and the SSIM was 0.92. The Detail Enhancement Factor (DEF) was 0.88, and the high-frequency component recovery (HFCR) was 0.09, which was at an excellent level compared to other style transfer methods and detail enhancement strategies. The results of this study indicate that the personalized poster generation model based on Pix2Pix exhibits significant advantages in style transfer and detail enhancement, effectively improving the quality of generated images and meeting users' personalized needs.<\/jats:p>","DOI":"10.1177\/14727978241307472","type":"journal-article","created":{"date-parts":[[2025,4,29]],"date-time":"2025-04-29T03:15:24Z","timestamp":1745896524000},"page":"1938-1950","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":1,"title":["Using Pix2Pix conditional generative adversarial networks to generate personalized poster content: Style transfer and detail enhancement"],"prefix":"10.1177","volume":"25","author":[{"given":"Rui","family":"Tian","sequence":"first","affiliation":[{"name":"State Grid Wuhan City Donghu New Technology Development Zone Electric Power Supply Company, Wuhan, Hubei, China"}]},{"given":"Xiao","family":"Li","sequence":"additional","affiliation":[{"name":"State Grid Wuhan City Donghu New Technology Development Zone Electric Power Supply Company, Wuhan, Hubei, China"}]},{"given":"Wenxi","family":"Li","sequence":"additional","affiliation":[{"name":"State Grid Wuhan City Donghu New Technology Development Zone Electric Power Supply Company, Wuhan, Hubei, China"}]},{"given":"Guojihong","family":"Li","sequence":"additional","affiliation":[{"name":"State Grid Wuhan City Donghu New Technology Development Zone Electric Power Supply Company, Wuhan, Hubei, China"}]},{"given":"Keer","family":"Chen","sequence":"additional","affiliation":[{"name":"State Grid Wuhan City Donghu New Technology Development Zone Electric Power Supply Company, Wuhan, Hubei, China"}]},{"given":"Hanwei","family":"Dai","sequence":"additional","affiliation":[{"name":"State Grid Wuhan City Donghu New Technology Development Zone Electric Power Supply Company, Wuhan, Hubei, China"}]}],"member":"179","published-online":{"date-parts":[[2024,12,10]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555437"},{"issue":"1","key":"e_1_3_2_3_2","first-page":"198","article-title":"Graphic style transfer technology in multimedia communication: an application of deep residual adaptive networks in graphic design","volume":"16","author":"Zhang Z","year":"2024","unstructured":"Zhang Z, Wei F, Liang G, et al. Graphic style transfer technology in multimedia communication: an application of deep residual adaptive networks in graphic design. Int J Commun Network Inf Secur 2024; 16(1): 198\u2013210.","journal-title":"Int J Commun Network Inf Secur"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3143356"},{"key":"e_1_3_2_5_2","doi-asserted-by":"crossref","unstructured":"Siddigan TM Sudharsan G Ram AGS. Pixelforge: image revamping through text and image input using pix2pix. In: Advances in Networks Intelligence and Computing. Milton: Taylor & Francis Group pp. 555\u2013561.","DOI":"10.1201\/9781003430421-56"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10043-022-00728-1"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jestch.2024.101827"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1162\/coli_a_00426"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-022-13596-3"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-022-12646-y"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.3390\/app13020867"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1049\/ipr2.12342"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.1002\/hbm.26422"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.comcom.2023.07.027"},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2022.06.009"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3167175"},{"key":"e_1_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1007\/s00259-021-05478-x"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1007\/s00500-022-06822-5"},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.3390\/rs14071574"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1002\/stc.3048"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3188433"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-021-11679-1"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.3390\/rs14102425"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-022-01719-z"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-023-17018-w"},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3228915"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2023.105053"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jksuci.2022.05.015"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.1145\/3592133"},{"key":"e_1_3_2_30_2","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555436"},{"key":"e_1_3_2_31_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2021.10.006"},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-022-16861-5"},{"key":"e_1_3_2_33_2","first-page":"1","article-title":"High resolution solar image generation using generative adversarial networks","volume":"11","author":"Dash A","year":"2022","unstructured":"Dash A, Ye J, Wang G, et al. High resolution solar image generation using generative adversarial networks. Ann Data Sci 2022; 11: 1\u201317.","journal-title":"Ann Data Sci"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.32604\/cmc.2023.041479"},{"key":"e_1_3_2_35_2","first-page":"131","article-title":"A review of image style transfer using generative adversarial networks techniques","volume":"23","author":"Zhang Y","year":"2024","unstructured":"Zhang Y. A review of image style transfer using generative adversarial networks techniques. Anal Metaphys 2024; 23: 131\u2013142.","journal-title":"Anal Metaphys"},{"key":"e_1_3_2_36_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-022-10777-x"},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1145\/3544903.3544906"},{"key":"e_1_3_2_38_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-022-03577-2"},{"key":"e_1_3_2_39_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10278-022-00696-7"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2021.116531"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jksuci.2020.03.007"}],"container-title":["Journal of Computational Methods in Sciences and Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.1177\/14727978241307472","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.1177\/14727978241307472","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.1177\/14727978241307472","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T16:32:26Z","timestamp":1771000346000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.1177\/14727978241307472"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,10]]},"references-count":40,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["10.1177\/14727978241307472"],"URL":"https:\/\/doi.org\/10.1177\/14727978241307472","relation":{},"ISSN":["1472-7978","1875-8983"],"issn-type":[{"value":"1472-7978","type":"print"},{"value":"1875-8983","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,10]]}}}