{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T04:06:54Z","timestamp":1745294814591,"version":"3.40.4"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031901669","type":"print"},{"value":"9783031901676","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-90167-6_15","type":"book-chapter","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T02:15:04Z","timestamp":1745288104000},"page":"218-232","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Exploring Bridges Between Algorithmic and\u00a0AI-Generated Art"],"prefix":"10.1007","author":[{"given":"Jiaqi","family":"Wu","sequence":"first","affiliation":[]},{"given":"Eytan","family":"Adar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,20]]},"reference":[{"key":"15_CR1","unstructured":"Ahn, N., et al.: Dreamstyler: Paint by style inversion with text-to-image diffusion models. arXiv preprint arXiv:2309.06933 (2023)"},{"key":"15_CR2","doi-asserted-by":"crossref","unstructured":"Angert, T., Suzara, M.I., Han, J., Pondoc, C.L., Subramonyam, H.: Spellburst: a node-based interface for exploratory creative coding with natural language prompts. In: UIST (2023)","DOI":"10.1145\/3586183.3606719"},{"key":"15_CR3","doi-asserted-by":"crossref","unstructured":"Barile, P., Ciesielski, V., Berry, M., Trist, K.: Animated drawings rendered by genetic programming. In: Proceedings of the 11th Annual Conference on Genetic and Evolutionary Computation, pp. 939\u2013946 (2009)","DOI":"10.1145\/1569901.1570030"},{"key":"15_CR4","doi-asserted-by":"crossref","unstructured":"Batista, J.E., Garrow, F., Huesca-Spairani, C., Martins, T.: Evoboard: geoboard-inspired evolved typefonts. In: International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar), pp. 17\u201332. Springer (2024)","DOI":"10.1007\/978-3-031-56992-0_2"},{"key":"15_CR5","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"325","DOI":"10.1007\/3-540-44617-6_31","volume-title":"Cognitive Technology: Instruments of Mind","author":"AF Blackwell","year":"2001","unstructured":"Blackwell, A.F., Britton, C., Cox, A., Green, T., Gurr, C., Kadoda, G., Kutar, M.S., Loomes, M., Nehaniv, C.L., Petre, M., Roast, C., Roe, C., Wong, A., Young, R.M.: Cognitive dimensions of notations: design tools for cognitive technology. In: Beynon, M., Nehaniv, C.L., Dautenhahn, K. (eds.) CT 2001. LNCS (LNAI), vol. 2117, pp. 325\u2013341. Springer, Heidelberg (2001). https:\/\/doi.org\/10.1007\/3-540-44617-6_31"},{"key":"15_CR6","unstructured":"Gal, R., et al.: An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618 (2022)"},{"key":"15_CR7","unstructured":"Galanter, P.: What is generative art? Complexity theory as a context for art theory (2003). https:\/\/api.semanticscholar.org\/CorpusID:2151469. Accessed 22 Jan 2025"},{"key":"15_CR8","unstructured":"Goodchild, A.: What is generative art? (2022). https:\/\/www.amygoodchild.com\/blog\/what-is-generative-art. Accessed 24 Jan 2025"},{"key":"15_CR9","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models (2020)"},{"key":"15_CR10","unstructured":"Jeong, J., Kim, J., Choi, Y., Lee, G., Uh, Y.: Visual style prompting with swapping self-attention (2024)"},{"key":"15_CR11","unstructured":"Liu, V., Long, T., Raw, N., Chilton, L.: Generative disco: Text-to-video generation for music visualization. arXiv preprint arXiv:2304.08551 (2023)"},{"key":"15_CR12","unstructured":"Luo, S., Tan, Y., Huang, L., Li, J., Zhao, H.: Latent consistency models: Synthesizing high-resolution images with few-step inference. arXiv preprint arXiv:2310.04378 (2023)"},{"key":"15_CR13","unstructured":"Meng, C., et al.: Sdedit: Guided image synthesis and editing with stochastic differential equations. arXiv preprint arXiv:2108.01073 (2021)"},{"key":"15_CR14","unstructured":"Peppler, K., Kafai, Y.: Creative coding: programming for personal expression, vol. 314 (2005). Accessed 30 Aug 2008"},{"key":"15_CR15","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"15_CR16","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., Aberman, K.: Dreambooth: fine tuning text-to-image diffusion models for subject-driven generation (2022)","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"15_CR17","volume-title":"The Nature of Code: Simulating Natural Systems with JavaScript","author":"D Shiffman","year":"2024","unstructured":"Shiffman, D.: The Nature of Code: Simulating Natural Systems with JavaScript. No Starch Press, San Francisco (2024)"},{"issue":"1","key":"15_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s13731-017-0072-1","volume":"6","author":"M Tempel","year":"2017","unstructured":"Tempel, M.: Generative art for all. J. Innov. Entrepreneurship 6(1), 1\u201314 (2017). https:\/\/doi.org\/10.1186\/s13731-017-0072-1","journal-title":"J. Innov. Entrepreneurship"},{"key":"15_CR19","unstructured":"Vaswani, A.: Attention is all you need. In: Advances in Neural Information Processing Systems (2017)"},{"key":"15_CR20","unstructured":"Wang, H., Wang, Q., Bai, X., Qin, Z., Chen, A.: Instantstyle: Free lunch towards style-preserving in text-to-image generation. arXiv preprint arXiv:2404.02733 (2024)"},{"key":"15_CR21","doi-asserted-by":"crossref","unstructured":"Wieluch, S., Schwenker, F.: Patternportrait: draw me like one of your scribbles. In: International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar), pp. 389\u2013400. Springer (2024)","DOI":"10.1007\/978-3-031-56992-0_25"},{"key":"15_CR22","doi-asserted-by":"crossref","unstructured":"Wiguna, I.P., Zen, A.P., Yuningsih, C.R.: Painting with algorithms: the potential for using the p5. js programming language for new media artist. In: Embracing the Future: Creative Industries for Environment and Advanced Society 5.0 in a Post-Pandemic Era, pp. 271\u2013275. Routledge (2022)","DOI":"10.1201\/9781003263135-55"},{"key":"15_CR23","doi-asserted-by":"crossref","unstructured":"Wu, T.: Saliency-aware generative art. In: Proceedings of the 2018 10th International Conference on Machine Learning and Computing, pp. 198\u2013202 (2018)","DOI":"10.1145\/3195106.3195143"},{"key":"15_CR24","unstructured":"Ye, H., Zhang, J., Liu, S., Han, X., Yang, W.: IP-adapter: text compatible image prompt adapter for text-to-image diffusion models (2023)"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Music, Sound, Art and Design"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-90167-6_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T02:15:18Z","timestamp":1745288118000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-90167-6_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031901669","9783031901676"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-90167-6_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"20 April 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EvoMUSART","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Trieste","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 April 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 April 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"evomusart2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.evostar.org\/2025\/evomusart\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}