{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T17:18:44Z","timestamp":1774027124886,"version":"3.50.1"},"reference-count":47,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["T2350005"],"award-info":[{"award-number":["T2350005"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["T2541041"],"award-info":[{"award-number":["T2541041"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Knowledge-Based Systems"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1016\/j.knosys.2026.115354","type":"journal-article","created":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T15:44:33Z","timestamp":1768751073000},"page":"115354","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Fontify : One-shot font generation via in-context learning"],"prefix":"10.1016","volume":"336","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-5835-8186","authenticated-orcid":false,"given":"Ying","family":"Xu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4415-0698","authenticated-orcid":false,"given":"Xiangwei","family":"Zhu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4052-1006","authenticated-orcid":false,"given":"Songyuan","family":"Li","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.knosys.2026.115354_bib0001","unstructured":"Y. Tian, Zi2zi: Master Chinese Calligraphy with Conditional Adversarial Networks, 2017, (https:\/\/github.com\/kaonashi-tyc\/zi2zi). GitHub repository."},{"key":"10.1016\/j.knosys.2026.115354_bib0002","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"10551","article-title":"Few-shot unsupervised image-to-image translation","author":"Liu","year":"2019"},{"key":"10.1016\/j.knosys.2026.115354_bib0003","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5130","article-title":"DG-font: deformable generative networks for unsupervised font generation","author":"Xie","year":"2021"},{"key":"10.1016\/j.knosys.2026.115354_bib0004","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"2393","article-title":"Few-shot font generation with localized style representations and factorization","volume":"35","author":"Park","year":"2021"},{"key":"10.1016\/j.knosys.2026.115354_bib0005","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"1436","article-title":"Multiple heads are better than one: few-shot font generation with multiple localized experts","author":"Park","year":"2021"},{"issue":"6","key":"10.1016\/j.knosys.2026.115354_bib0006","first-page":"1","article-title":"Deepvecfont: synthesizing high-quality vector fonts via dual-modality learning","volume":"40","author":"Wang","year":"2021","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"10.1016\/j.knosys.2026.115354_bib0007","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"3774","article-title":"Deepcallifont: few-shot chinese calligraphy font synthesis by integrating dual-modality generative models","volume":"38","author":"Liu","year":"2024"},{"key":"10.1016\/j.knosys.2026.115354_bib0008","doi-asserted-by":"crossref","first-page":"25005","DOI":"10.52202\/068431-1813","article-title":"Visual prompting via image inpainting","volume":"35","author":"Bar","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0009","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"6830","article-title":"Images speak in images: a generalist painter for in-context visual learning","author":"Wang","year":"2023"},{"key":"10.1016\/j.knosys.2026.115354_bib0010","series-title":"SIGGRAPH Asia 2017 Technical Briefs","first-page":"1","article-title":"DCFont: an end-To-end deep chinese font generation system","author":"Jiang","year":"2017"},{"key":"10.1016\/j.knosys.2026.115354_bib0011","series-title":"BMVC","first-page":"290","article-title":"Chinese handwriting imitation with hierarchical generative adversarial network","author":"Chang","year":"2018"},{"key":"10.1016\/j.knosys.2026.115354_bib0012","unstructured":"Y. Tian, Rewrite: Neural Style Transfer for Chinese Fonts, 2016, (https:\/\/github.com\/kaonashi-tyc\/Rewrite). GitHub repository."},{"key":"10.1016\/j.knosys.2026.115354_bib0013","series-title":"2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR)","first-page":"1095","article-title":"Auto-encoder guided GAN for chinese calligraphy synthesis","author":"Lyu","year":"2017"},{"key":"10.1016\/j.knosys.2026.115354_bib0014","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2021.107334","article-title":"CS-GAN: cross-structure generative adversarial networks for chinese calligraphy translation","volume":"229","author":"Xiao","year":"2021","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0015","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"1125","article-title":"Image-to-image translation with conditional adversarial networks","author":"Isola","year":"2017"},{"key":"10.1016\/j.knosys.2026.115354_bib0016","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"8789","article-title":"Stargan: unified generative adversarial networks for multi-domain image-to-image translation","author":"Choi","year":"2018"},{"key":"10.1016\/j.knosys.2026.115354_bib0017","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"2107","article-title":"Learning from simulated and unsupervised images through adversarial training","author":"Shrivastava","year":"2017"},{"key":"10.1016\/j.knosys.2026.115354_bib0018","series-title":"Proceedings of the IEEE International Conference on Computer Vision","first-page":"2223","article-title":"Unpaired image-to-image translation using cycle-consistent adversarial networks","author":"Zhu","year":"2017"},{"key":"10.1016\/j.knosys.2026.115354_bib0019","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"8798","article-title":"High-resolution image synthesis and semantic manipulation with conditional gans","author":"Wang","year":"2018"},{"key":"10.1016\/j.knosys.2026.115354_bib0020","first-page":"1","article-title":"Artistic glyph image synthesis via one-Stage few-Shot learning","author":"Gao","year":"2019","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"10.1016\/j.knosys.2026.115354_bib0021","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"1858","article-title":"CF-font: content fusion for few-shot font generation","author":"Wang","year":"2023"},{"key":"10.1016\/j.knosys.2026.115354_bib0022","series-title":"2025 8th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE)","first-page":"684","article-title":"CA-font: a chinese font generation model based on font style transfer","author":"Wan","year":"2025"},{"key":"10.1016\/j.knosys.2026.115354_bib0023","series-title":"Proceedings of the 33rd ACM International Conference on Multimedia","first-page":"6644","article-title":"DA-font: few-shot font generation via dual-attention hybrid integration","author":"Chen","year":"2025"},{"key":"10.1016\/j.knosys.2026.115354_bib0024","series-title":"GC-Font: Few-Shot Font Generation via Global Contextual Feature Modelling","author":"Chen","year":"2025"},{"key":"10.1016\/j.knosys.2026.115354_bib0025","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"6603","article-title":"Fontdiffuser: one-shot font generation via denoising diffusion with multi-scale content aggregation and style contrastive learning","volume":"38","author":"Yang","year":"2024"},{"key":"10.1016\/j.knosys.2026.115354_bib0026","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"6892","article-title":"Generate like experts: multi-stage font generation by incorporating font transfer process into diffusion models","author":"Fu","year":"2024"},{"key":"10.1016\/j.knosys.2026.115354_bib0027","doi-asserted-by":"crossref","unstructured":"B. Lu, T. Zhao, Y. Liu, G. Liang, G. Wang, C. Wang, G. Zhan, F.B. Khalid, Difffontseed: a structure-aware diffusion-based framework for few-shot chinese font generation, Available at SSRN 5382851 (2025).","DOI":"10.2139\/ssrn.5382851"},{"key":"10.1016\/j.knosys.2026.115354_bib0028","series-title":"Asian Conference on Pattern Recognition","first-page":"324","article-title":"FLUX-font: stylized font generation based on fine-tuning pre-trained diffusion transformer","author":"Li","year":"2025"},{"key":"10.1016\/j.knosys.2026.115354_bib0029","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2025.114600","article-title":"SGD-font: style and glyph decoupling for one-shot font generation","author":"Li","year":"2025","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0030","doi-asserted-by":"crossref","unstructured":"D. Sun, T. Ren, C. Li, H. Su, J. Zhu, Learning to write stylized chinese characters by reading a handful of examples, (2017). arXiv: 1712.06424.","DOI":"10.24963\/ijcai.2018\/128"},{"key":"10.1016\/j.knosys.2026.115354_bib0031","unstructured":"S.-J. Wu, C.-Y. Yang, J.Y.-j. Hsu, Calligan: style and structure-aware chinese calligraphy character generator, (2020). arXiv: 2005.12500."},{"key":"10.1016\/j.knosys.2026.115354_bib0032","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"3270","article-title":"Strokegan: reducing mode collapse in chinese font generation via stroke encoding","volume":"35","author":"Zeng","year":"2021"},{"key":"10.1016\/j.knosys.2026.115354_bib0033","series-title":"European Conference on Computer Vision","first-page":"735","article-title":"Few-shot compositional font generation with dual memory","author":"Cha","year":"2020"},{"key":"10.1016\/j.knosys.2026.115354_bib0034","doi-asserted-by":"crossref","first-page":"14177","DOI":"10.52202\/079017-0454","article-title":"IF-font: ideographic description sequence-following font generation","volume":"37","author":"Chen","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0035","doi-asserted-by":"crossref","DOI":"10.1016\/j.dsp.2025.105305","article-title":"FourcornerGAN: glyph formation augmentation for unpaired chinese font generation","author":"Wang","year":"2025","journal-title":"Digit. Signal Process."},{"key":"10.1016\/j.knosys.2026.115354_bib0036","series-title":"Computer Graphics Forum","first-page":"567","article-title":"FontRNN: generating large-scale chinese fonts via recurrent neural network","volume":"38","author":"Tang","year":"2019"},{"key":"10.1016\/j.knosys.2026.115354_bib0037","series-title":"Computer Graphics Forum","first-page":"141","article-title":"Write like you: synthesizing your cursive online chinese handwriting via metric-based meta learning","volume":"40","author":"Tang","year":"2021"},{"key":"10.1016\/j.knosys.2026.115354_bib0038","series-title":"ACM SIGGRAPH 2024 Conference Papers","first-page":"1","article-title":"QT-font: high-efficiency font synthesis via quadtree-based diffusion models","author":"Liu","year":"2024"},{"key":"10.1016\/j.knosys.2026.115354_bib0039","article-title":"Few-Shot font generation via stroke prompt and hierarchical representation learning","author":"Zeng","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.knosys.2026.115354_bib0040","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0041","first-page":"52692","article-title":"Spae: semantic pyramid autoencoder for multimodal generation with frozen llms","volume":"36","author":"Yu","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0042","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.111382","article-title":"Generating high-quality texture via panoramic feature aggregation for large mask inpainting","volume":"286","author":"Wang","year":"2024","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0043","article-title":"Faster r-cnn: towards real-time object detection with region proposal networks","volume":"28","author":"Ren","year":"2015","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115354_bib0044","series-title":"Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"433","article-title":"Few-shot font style transfer between different languages","author":"Li","year":"2021"},{"key":"10.1016\/j.knosys.2026.115354_bib0045","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"586","article-title":"The unreasonable effectiveness of deep features as a perceptual metric","author":"Zhang","year":"2018"},{"key":"10.1016\/j.knosys.2026.115354_bib0046","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"4681","article-title":"Photo-realistic single image super-resolution using a generative adversarial network","author":"Ledig","year":"2017"},{"key":"10.1016\/j.knosys.2026.115354_bib0047","series-title":"Proceedings of the IEEE International Conference on Computer Vision","first-page":"4491","article-title":"Enhancenet: single image super-resolution through automated texture synthesis","author":"Sajjadi","year":"2017"}],"container-title":["Knowledge-Based Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126000973?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126000973?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T16:30:22Z","timestamp":1774024222000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0950705126000973"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3]]},"references-count":47,"alternative-id":["S0950705126000973"],"URL":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115354","relation":{},"ISSN":["0950-7051"],"issn-type":[{"value":"0950-7051","type":"print"}],"subject":[],"published":{"date-parts":[[2026,3]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Fontify : One-shot font generation via in-context learning","name":"articletitle","label":"Article Title"},{"value":"Knowledge-Based Systems","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115354","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"115354"}}