{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T00:49:15Z","timestamp":1774658955421,"version":"3.50.1"},"reference-count":68,"publisher":"Springer Science and Business Media LLC","issue":"35","license":[{"start":{"date-parts":[[2025,5,5]],"date-time":"2025-05-05T00:00:00Z","timestamp":1746403200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,5,5]],"date-time":"2025-05-05T00:00:00Z","timestamp":1746403200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-025-20874-3","type":"journal-article","created":{"date-parts":[[2025,5,5]],"date-time":"2025-05-05T04:45:40Z","timestamp":1746420340000},"page":"43805-43826","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Gorgeous: Creating narrative-driven makeup ideas via image prompts"],"prefix":"10.1007","volume":"84","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-3847-8175","authenticated-orcid":false,"given":"Jia Wei","family":"Sii","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7677-2865","authenticated-orcid":false,"given":"Chee Seng","family":"Chan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,5]]},"reference":[{"key":"20874_CR1","volume-title":"Face Paint: The Story of Makeup","author":"L Eldridge","year":"2015","unstructured":"Eldridge L (2015) Face Paint: The Story of Makeup. Abrams, New York City"},{"key":"20874_CR2","doi-asserted-by":"crossref","unstructured":"Yang C, He W, Xu Y, Gao Y (2022) Elegant: Exquisite and locally editable gan for makeup transfer. In: ECCV. Springer","DOI":"10.1007\/978-3-031-19787-1_42"},{"key":"20874_CR3","doi-asserted-by":"crossref","unstructured":"Sun Z, Chen Y, Xiong S (2022) Ssat: A symmetric semantic-aware transformer network for makeup transfer and removal. In: AAAI","DOI":"10.1609\/aaai.v36i2.20131"},{"key":"20874_CR4","doi-asserted-by":"crossref","unstructured":"Yan Q, Guo C, Zhao J, Dai Y, Loy CC, Li C (2023) Beautyrec: Robust, efficient, and component-specific makeup transfer. In: CVPR","DOI":"10.1109\/CVPRW59228.2023.00117"},{"key":"20874_CR5","doi-asserted-by":"crossref","unstructured":"Zhang Y, Wei L, Zhang Q, Song Y, Liu J, Li H, Tang X, Hu Y, Zhao H (2024) Stable-Makeup: When Real-World Makeup Transfer Meets Diffusion Model","DOI":"10.1145\/3721238.3730702"},{"key":"20874_CR6","unstructured":"Guo D, Sim T (2009) Digital face makeup by example. In: CVPR. IEEE"},{"key":"20874_CR7","unstructured":"Liu S, Ou X, Qian R, Wang W, Cao X (2016) Makeup like a superstar: Deep localized makeup transfer network. arXiv preprint arXiv:1604.07102"},{"key":"20874_CR8","doi-asserted-by":"crossref","unstructured":"Chang H, Lu J, Yu F, Finkelstein A (2018) Pairedcyclegan: Asymmetric style transfer for applying and removing makeup. In: CVPR","DOI":"10.1109\/CVPR.2018.00012"},{"key":"20874_CR9","doi-asserted-by":"crossref","unstructured":"Li T, Qian R, Dong C, Liu S, Yan Q, Zhu W, Lin L (2018) Beautygan: Instance-level facial makeup transfer with deep generative adversarial network. In: ACMMM","DOI":"10.1145\/3240508.3240618"},{"key":"20874_CR10","doi-asserted-by":"crossref","unstructured":"Chen H-J, Hui K-M, Wang S-Y, Tsao L-W, Shuai H-H, Cheng W-H (2019) Beautyglow: On-demand makeup transfer framework with reversible generative network. In: CVPR","DOI":"10.1109\/CVPR.2019.01028"},{"key":"20874_CR11","doi-asserted-by":"crossref","unstructured":"Deng H, Han C, Cai H, Han G, He S (2021) Spatially-invariant style-codes controlled makeup transfer. In: CVPR","DOI":"10.1109\/CVPR46437.2021.00648"},{"key":"20874_CR12","doi-asserted-by":"crossref","unstructured":"Jiang W, Liu S, Gao C, Cao J, He R, Feng J, Yan S (2020) Psgan: Pose and expression robust spatial-aware gan for customizable makeup transfer. In: CVPR","DOI":"10.1109\/CVPR42600.2020.00524"},{"key":"20874_CR13","doi-asserted-by":"crossref","unstructured":"Lyu Y, Dong J, Peng B, Wang W, Tan T (2021) Sogan: 3d-aware shadow and occlusion robust gan for makeup transfer. In: ACMMM","DOI":"10.1145\/3474085.3475531"},{"key":"20874_CR14","doi-asserted-by":"crossref","unstructured":"Nguyen T, Tran AT, Hoai M (2021) Lipstick ain\u2019t enough: beyond color matching for in-the-wild makeup transfer. In: CVPR","DOI":"10.1109\/CVPR46437.2021.01310"},{"key":"20874_CR15","doi-asserted-by":"crossref","unstructured":"Jin Q, Chen X, Jin M, Cheng Y, Shi R, Zheng Y, Zhu Y, Ni B (2024) Toward tiny and high-quality facial makeup with data amplify learning. arXiv preprint arXiv:2403.15033","DOI":"10.1007\/978-3-031-73383-3_20"},{"key":"20874_CR16","doi-asserted-by":"crossref","unstructured":"Nguyen TV, Liu L (2017) Smart mirror: Intelligent makeup recommendation and synthesis. In: ACMMM","DOI":"10.1145\/3123266.3127926"},{"key":"20874_CR17","doi-asserted-by":"crossref","unstructured":"Alashkar T, Jiang S, Wang S, Fu Y (2017) Examples-rules guided deep neural network for makeup recommendation. In: AAAI","DOI":"10.1609\/aaai.v31i1.10626"},{"key":"20874_CR18","doi-asserted-by":"crossref","unstructured":"Alashkar T, Jiang S, Fu Y (2017) Rule-based facial makeup recommendation system. In: FG","DOI":"10.1109\/FG.2017.47"},{"key":"20874_CR19","doi-asserted-by":"crossref","unstructured":"Perera P, Soysa E, De\u00a0Silva H, Tavarayan A, Gamage M, Weerasinghe K (2021) Virtual makeover and makeup recommendation based on personal trait analysis. In: ICAC","DOI":"10.1109\/ICAC54203.2021.9671133"},{"key":"20874_CR20","unstructured":"Gulati K, Verma G, Mohania M, Kundu A (2023) Beautifai-personalised occasion-based makeup recommendation. In: ACML"},{"key":"20874_CR21","doi-asserted-by":"crossref","unstructured":"Jing Y, Yang Y, Feng Z, Ye J, Yu Y, Song M (2020) Neural style transfer: A review. T-VCG","DOI":"10.1109\/TVCG.2019.2921336"},{"key":"20874_CR22","doi-asserted-by":"crossref","unstructured":"Cai Q, Ma M, Wang C, Li H (2023) Image neural style transfer: A review. Computers and Electrical Engineering","DOI":"10.1016\/j.compeleceng.2023.108723"},{"key":"20874_CR23","doi-asserted-by":"crossref","unstructured":"Li Y, Wang N, Liu J, Hou X (2017) Demystifying neural style transfer. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence. IJCAI\u201917","DOI":"10.24963\/ijcai.2017\/310"},{"key":"20874_CR24","doi-asserted-by":"crossref","unstructured":"Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: CVPR","DOI":"10.1109\/CVPR.2016.265"},{"key":"20874_CR25","doi-asserted-by":"crossref","unstructured":"Zhang W, Cao C, Chen S, Liu J, Tang X (2013) Style transfer via image component analysis. TMM","DOI":"10.1109\/TMM.2013.2265675"},{"key":"20874_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073683","volume":"36","author":"J Liao","year":"2017","unstructured":"Liao J, Yao Y, Yuan L, Hua G, Kang SB (2017) Visual attribute transfer through deep image analogy. ACM Transactions on Graphics (TOG). 36:1\u201315","journal-title":"ACM Transactions on Graphics (TOG)."},{"key":"20874_CR27","doi-asserted-by":"crossref","unstructured":"Gu S, Chen C, Liao J, Yuan L (2018) Arbitrary style transfer with deep feature reshuffle. In: CVPR","DOI":"10.1109\/CVPR.2018.00858"},{"key":"20874_CR28","unstructured":"Zhang Z, Liu Y, Han C, Guo T, Yao T, Mei T (2022) Generalized one-shot domain adaptation of generative adversarial networks. NeurIPS"},{"key":"20874_CR29","doi-asserted-by":"crossref","unstructured":"Kolkin N, Salavon J, Shakhnarovich G (2019) Style transfer by relaxed optimal transport and self-similarity. In: CVPR","DOI":"10.1109\/CVPR.2019.01029"},{"key":"20874_CR30","doi-asserted-by":"crossref","unstructured":"Deng Y, Tang F, Dong W, Sun W, Huang F, Xu C (2020) Arbitrary style transfer via multi-adaptation network. In: ACMMM","DOI":"10.1145\/3394171.3414015"},{"key":"20874_CR31","doi-asserted-by":"crossref","unstructured":"Yao Y, Ren J, Xie X, Liu W, Liu Y-J, Wang J (2019) Attention-aware multi-stroke style transfer. In: CVPR","DOI":"10.1109\/CVPR.2019.00156"},{"key":"20874_CR32","doi-asserted-by":"crossref","unstructured":"Liu S, Lin T, He D, Li F, Wang M, Li X, Sun Z, Li Q, Ding E (2021) Adaattn: Revisit attention mechanism in arbitrary neural style transfer. In: ICCV","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"20874_CR33","doi-asserted-by":"crossref","unstructured":"An J, Huang S, Song Y, Dou D, Liu W, Luo J (2021) Artflow: Unbiased image style transfer via reversible neural flows. In: CVPR","DOI":"10.1109\/CVPR46437.2021.00092"},{"key":"20874_CR34","unstructured":"Chen H, Wang Z, Zhang H, Zuo Z, Li A, Xing W, Lu D et al (2021) Artistic style transfer with internal-external learning and contrastive learning. In: NeurIPS"},{"key":"20874_CR35","doi-asserted-by":"crossref","unstructured":"Zhang Y, Tang F, Dong W, Huang H, Ma C, Lee T-Y, Xu C (2022) Domain enhanced arbitrary image style transfer via contrastive learning. In: SIGGRAPH","DOI":"10.1145\/3528233.3530736"},{"key":"20874_CR36","doi-asserted-by":"crossref","unstructured":"Wu X, Hu Z, Sheng L, Xu D (2021) Styleformer: Real-time arbitrary style transfer via parametric style composition. In: ICCV","DOI":"10.1109\/ICCV48922.2021.01435"},{"key":"20874_CR37","doi-asserted-by":"crossref","unstructured":"Deng Y, Tang F, Dong W, Ma C, Pan X, Wang L, Xu C (2022) Stytr2: Image style transfer with transformers. In: CVPR","DOI":"10.1109\/CVPR52688.2022.01104"},{"key":"20874_CR38","doi-asserted-by":"crossref","unstructured":"Cheng B, Liu Z, Peng Y, Lin Y (2023) General image-to-image translation with one-shot image guidance. In: ICCV","DOI":"10.1109\/ICCV51070.2023.02078"},{"key":"20874_CR39","doi-asserted-by":"crossref","unstructured":"Zhang Y, Huang N, Tang F, Huang H, Ma C, Dong W, Xu C (2023) Inversion-based style transfer with diffusion models. In: CVPR","DOI":"10.1109\/CVPR52729.2023.00978"},{"key":"20874_CR40","doi-asserted-by":"crossref","unstructured":"Huo J, Liu X, Li W, Gao Y, Yin H, Luo J (2022) Cast: Learning both geometric and texture style transfers for effective caricature generation. TIP","DOI":"10.1109\/TIP.2022.3154238"},{"key":"20874_CR41","unstructured":"Ramesh A, Dhariwal P, Nichol A, Chu C, Chen M (2022) Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125"},{"key":"20874_CR42","unstructured":"Balaji Y, Nah S, Huang X, Vahdat A, Song J, Kreis K, Aittala M, Aila T, Laine S, Catanzaro B et al (2022) ediffi: Text-to-image diffusion models with an ensemble of expert denoisers. arXiv preprint arXiv:2211.01324"},{"key":"20874_CR43","unstructured":"Ding M, Zheng W, Hong W, Tang J (2022) Cogview2: Faster and better text-to-image generation via hierarchical transformers. In: NeurIPS"},{"key":"20874_CR44","unstructured":"Nichol A, Dhariwal P, Ramesh A, Shyam P, Mishkin P, McGrew B, Sutskever I, Chen M (2021) Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741"},{"key":"20874_CR45","doi-asserted-by":"crossref","unstructured":"Saharia C, Chan W, Saxena S, Li L, Whang J, Denton EL, Ghasemipour K, Gontijo\u00a0Lopes R, Karagol\u00a0Ayan B, Salimans T, et al. (2022) Photorealistic text-to-image diffusion models with deep language understanding. In: NeurIPS","DOI":"10.1145\/3528233.3530757"},{"key":"20874_CR46","unstructured":"Ding M, Yang Z, Hong W, Zheng W, Zhou C, Yin D, Lin J, Zou X, Shao Z, Yang H et al (2021) Cogview: Mastering text-to-image generation via transformers. In: NeurIPS"},{"key":"20874_CR47","doi-asserted-by":"crossref","unstructured":"Zhang L, Rao A, Agrawala M (2023) Adding conditional control to text-to-image diffusion models. ICCV","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"20874_CR48","unstructured":"Dhariwal P, Nichol A (2021) Diffusion models beat gans on image synthesis. In: NeurIPS"},{"key":"20874_CR49","doi-asserted-by":"crossref","unstructured":"Brooks T, Holynski A, Efros AA (2023) Instructpix2pix: Learning to follow image editing instructions. In: CVPR","DOI":"10.1109\/CVPR52729.2023.01764"},{"key":"20874_CR50","unstructured":"Gal R, Alaluf Y, Atzmon Y, Patashnik O, Bermano AH, Chechik G, Cohen-Or D (2022) An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618"},{"key":"20874_CR51","doi-asserted-by":"crossref","unstructured":"Ruiz N, Li Y, Jampani V, Pritch Y, Rubinstein M, Aberman K (2023) Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In: CVPR","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"20874_CR52","doi-asserted-by":"crossref","unstructured":"Tumanyan N, Geyer M, Bagon S, Dekel T (2023) Plug-and-play diffusion features for text-driven image-to-image translation. In: CVPR","DOI":"10.1109\/CVPR52729.2023.00191"},{"key":"20874_CR53","doi-asserted-by":"crossref","unstructured":"Kawar B, Zada S, Lang O, Tov O, Chang H-T, Dekel T, Mosseri I, Irani M (2023) Imagic: Text-based real image editing with diffusion models. In: CVPR","DOI":"10.1109\/CVPR52729.2023.00582"},{"key":"20874_CR54","doi-asserted-by":"crossref","unstructured":"Kim G, Kwon T, Ye J-C (2022) Diffusionclip: Text-guided diffusion models for robust image manipulation. In: CVPR","DOI":"10.1109\/CVPR52688.2022.00246"},{"key":"20874_CR55","unstructured":"Podell D, English Z, Lacey K, Blattmann A, Dockhorn T, M\u00fcller J, Penna J, Rombach R (2023) Sdxl: Improving latent diffusion models for high-resolution image synthesis. arXiv preprint arXiv:2307.01952"},{"key":"20874_CR56","doi-asserted-by":"crossref","unstructured":"Parmar G, Singh KK, Zhang R, Li Y, Lu J, Zhu J-Y (2023) Zero-shot image-to-image translation. In: SIGGRAPH","DOI":"10.1145\/3588432.3591513"},{"key":"20874_CR57","doi-asserted-by":"crossref","unstructured":"Avrahami O, Lischinski D, Fried O (2022) Blended diffusion for text-driven editing of natural images. In: CVPR","DOI":"10.1109\/CVPR52688.2022.01767"},{"key":"20874_CR58","unstructured":"Meng C, He Y, Song Y, Song J, Wu J, Zhu J-Y, Ermon S (2021) Sdedit: Guided image synthesis and editing with stochastic differential equations. In: ICLR"},{"key":"20874_CR59","unstructured":"Ye H, Zhang J, Liu S, Han X, Yang W (2023) Ip-adapter: Text compatible image prompt adapter for text-to-image diffusion models. arXiv preprint arXiv:2308.06721"},{"key":"20874_CR60","doi-asserted-by":"crossref","unstructured":"Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B (2022) High-resolution image synthesis with latent diffusion models. In: CVPR","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"20874_CR61","unstructured":"Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. In: ICML"},{"key":"20874_CR62","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: MICCAI","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"20874_CR63","doi-asserted-by":"crossref","unstructured":"Gu Q, Wang G, Chiu MT, Tai Y-W, Tang C-K (2019) Ladn: Local adversarial disentangling network for facial makeup and de-makeup. In: CVPR","DOI":"10.1109\/ICCV.2019.01058"},{"key":"20874_CR64","doi-asserted-by":"crossref","unstructured":"Yu C, Wang J, Peng C, Gao C, Yu G, Sang N (2018) Bisenet: Bilateral segmentation network for real-time semantic segmentation. In: ECCV","DOI":"10.1007\/978-3-030-01261-8_20"},{"key":"20874_CR65","doi-asserted-by":"crossref","unstructured":"Somepalli G, Gupta A, Gupta K, Palta S, Goldblum M, Geiping J, Shrivastava A, Goldstein T (2024) Measuring style similarity in diffusion models. arXiv preprint arXiv:2404.01292","DOI":"10.1007\/978-3-031-72848-8_9"},{"key":"20874_CR66","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS"},{"key":"20874_CR67","unstructured":"Wang H, Xing P, Huang R, Ai H, Wang Q, Bai X (2024) InstantStyle-Plus: Style Transfer with Content-Preserving in Text-to-Image Generation. https:\/\/arxiv.org\/abs\/2407.00788"},{"key":"20874_CR68","unstructured":"Fu S, Tamir N, Sundaram S, Chai L, Zhang R, Dekel T, Isola P (2023) Dreamsim: Learning new dimensions of human visual similarity using synthetic data. In: NeurIPS"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-025-20874-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-025-20874-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-025-20874-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T10:50:34Z","timestamp":1761389434000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-025-20874-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,5]]},"references-count":68,"journal-issue":{"issue":"35","published-online":{"date-parts":[[2025,10]]}},"alternative-id":["20874"],"URL":"https:\/\/doi.org\/10.1007\/s11042-025-20874-3","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,5]]},"assertion":[{"value":"5 December 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 April 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 April 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 May 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"No competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}