{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,17]],"date-time":"2026-01-17T10:59:18Z","timestamp":1768647558266,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":33,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556786","type":"print"},{"value":"9789819556793","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5679-3_23","type":"book-chapter","created":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T18:36:45Z","timestamp":1768329405000},"page":"330-342","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["StrategyAdapter: One-Shot Learning for\u00a0Unseen-Domain Procedural Sequence Generation"],"prefix":"10.1007","author":[{"given":"Yingying","family":"Sun","sequence":"first","affiliation":[]},{"given":"Zhiguang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Nong","family":"Xiao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,14]]},"reference":[{"key":"23_CR1","unstructured":"Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., et\u00a0al.: GPT-4 technical report. arXiv preprint arXiv:2303.08774 (2023)"},{"key":"23_CR2","doi-asserted-by":"crossref","unstructured":"Chen, B., Wang, Y., et\u00a0al.: Inverse painting: reconstructing the painting process. In: SIGGRAPH Asia 2024 Conference Papers, pp. 1\u201311 (2024)","DOI":"10.1145\/3680528.3687574"},{"key":"23_CR3","unstructured":"Esser, P., Kulal, S., et\u00a0al.: Scaling rectified flow transformers for high-resolution image synthesis. In: Forty-First International Conference on Machine Learning (2024)"},{"key":"23_CR4","first-page":"5207","volume":"35","author":"K Frans","year":"2022","unstructured":"Frans, K., Soros, L., Witkowski, O.: CLIPDraw: exploring text-to-drawing synthesis through language-image encoders. Adv. Neural. Inf. Process. Syst. 35, 5207\u20135218 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"23_CR5","unstructured":"Ha, D., Eck, D.: A neural representation of sketch drawings. arXiv preprint arXiv:1704.03477 (2017)"},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Haeberli, P.: Paint by numbers: abstract image representations. In: Proceedings of the 17th Annual Conference on Computer Graphics and Interactive Techniques, pp. 207\u2013214 (1990)","DOI":"10.1145\/97879.97902"},{"key":"23_CR7","doi-asserted-by":"crossref","unstructured":"Hertzmann, A.: A survey of stroke-based rendering. Institute of Electrical and Electronics Engineers (2003)","DOI":"10.1109\/MCG.2003.1210867"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Hessel, J., Holtzman, A., Forbes, M., Bras, R.L., Choi, Y.: CLIPScore: a reference-free evaluation metric for image captioning. In: EMNLP (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.595"},{"key":"23_CR9","unstructured":"Hu, E.J., Shen, Y., et\u00a0al.: LoRA: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)"},{"key":"23_CR10","unstructured":"Huang, Z., Wei, Y., et\u00a0al.: Denoising diffusion probabilistic models are optimally adaptive to unknown low dimensionality. arXiv preprint arXiv:2410.18784 (2024)"},{"key":"23_CR11","doi-asserted-by":"crossref","unstructured":"Kotovenko, D., et\u00a0al.: Rethinking style transfer: from pixels to parameterized brushstrokes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12196\u201312205 (2021)","DOI":"10.1109\/CVPR46437.2021.01202"},{"key":"23_CR12","unstructured":"Labs, B.F.: FLUX: official inference repository for FLUX.1 models (2024). https:\/\/github.com\/BlackForestLabs\/flux. Accessed 12 Nov 2024"},{"key":"23_CR13","doi-asserted-by":"crossref","unstructured":"Litwinowicz, P.: Processing images and video for an impressionist effect. In: Proceedings of the 24th Annual Conference on Computer Graphics and Interactive Techniques, pp. 407\u2013414 (1997)","DOI":"10.1145\/258734.258893"},{"key":"23_CR14","unstructured":"MidJourney: AI-generated image of procedural sequence (2025). Version 5 - https:\/\/www.midjourney.com"},{"key":"23_CR15","unstructured":"Mishchenko, K., Defazio, A.: Prodigy: an expeditiously adaptive parameter-free learner. arXiv preprint arXiv:2306.06101 (2023)"},{"key":"23_CR16","unstructured":"Nakano, R.: Neural Painters: a learned differentiable constraint for generating brushstroke paintings. arXiv preprint arXiv:1904.08410 (2019)"},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Peebles, W., et\u00a0al.: Scalable diffusion models with transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4195\u20134205 (2023)","DOI":"10.1109\/ICCV51070.2023.00387"},{"key":"23_CR18","unstructured":"Radford, A., Kim, J.W., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., et\u00a0al.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"23_CR20","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Li, Y., et\u00a0al.: DreamBooth: fine tuning text-to-image diffusion models for subject-driven generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22500\u201322510 (2023)","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"23_CR21","unstructured":"Shazeer, N., Mirhoseini, A., et\u00a0al.: Outrageously large neural networks: the sparsely-gated mixture-of-experts layer. arXiv preprint arXiv:1701.06538 (2017)"},{"key":"23_CR22","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502 (2020)"},{"key":"23_CR23","doi-asserted-by":"crossref","unstructured":"Song, Y.: CLIPTexture: text-driven texture synthesis. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 5468\u20135476 (2022)","DOI":"10.1145\/3503161.3548146"},{"key":"23_CR24","unstructured":"Song, Y., Huang, S., et\u00a0al.: ProcessPainter: learn painting process from sequence data. arXiv preprint arXiv:2406.06062 (2024)"},{"key":"23_CR25","unstructured":"Song, Y., Liu, C., et\u00a0al.: MakeAnything: harnessing diffusion transformers for multi-domain procedural sequence generation. arXiv preprint arXiv:2502.01572 (2025)"},{"key":"23_CR26","doi-asserted-by":"crossref","unstructured":"Song, Y., et\u00a0al.: CLIPVG: text-guided image manipulation using differentiable vector graphics. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a037, pp. 2312\u20132320 (2023)","DOI":"10.1609\/aaai.v37i2.25326"},{"key":"23_CR27","unstructured":"InstantX Team: InstantX FLUX.1-dev IP-adapter page (2024)"},{"key":"23_CR28","unstructured":"Paints-Undo Team: Paints-undo GitHub page (2024)"},{"key":"23_CR29","unstructured":"Wan, C., Luo, X., et\u00a0al.: GRID: visual layout generation. arXiv preprint arXiv:2412.10718 (2024)"},{"issue":"5","key":"23_CR30","doi-asserted-by":"publisher","first-page":"1134","DOI":"10.1587\/transinf.E96.D.1134","volume":"96","author":"N Xie","year":"2013","unstructured":"Xie, N., Hachiya, H., Sugiyama, M.: Artist agent: a reinforcement learning approach to automatic stroke generation in oriental ink painting. IEICE Trans. Inf. Syst. 96(5), 1134\u20131144 (2013)","journal-title":"IEICE Trans. Inf. Syst."},{"key":"23_CR31","unstructured":"Ye, H., Zhang, J., Liu, S., et\u00a0al.: IP-Adapter: text compatible image prompt adapter for text-to-image diffusion models (2023)"},{"key":"23_CR32","doi-asserted-by":"crossref","unstructured":"Zhai, X., Mustafa, B., et\u00a0al.: Sigmoid loss for language image pre-training. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 11975\u201311986 (2023)","DOI":"10.1109\/ICCV51070.2023.01100"},{"key":"23_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Song, Y., et\u00a0al.: SSR-Encoder: encoding selective subject representation for subject-driven generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8069\u20138078 (2024)","DOI":"10.1109\/CVPR52733.2024.00771"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5679-3_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T18:36:48Z","timestamp":1768329408000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5679-3_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556786","9789819556793"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5679-3_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"14 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}