{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T04:06:55Z","timestamp":1745294815033,"version":"3.40.4"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031901669","type":"print"},{"value":"9783031901676","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-90167-6_17","type":"book-chapter","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T02:15:06Z","timestamp":1745288106000},"page":"249-260","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["All YIN No YANG: Geometric Abstraction of Oil Paintings with Trained Models, Noise and Self-reference"],"prefix":"10.1007","author":[{"given":"Lu\u00eds","family":"Arandas","sequence":"first","affiliation":[]},{"given":"Iulia","family":"Ionescu","sequence":"additional","affiliation":[]},{"given":"Murad","family":"Khan","sequence":"additional","affiliation":[]},{"given":"Mick","family":"Grierson","sequence":"additional","affiliation":[]},{"given":"Miguel","family":"Carvalhais","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,20]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Arandas, L., Grierson, M., Carvalhais, M.: Computing short films using language-guided diffusion and vocoding through virtual timelines of summaries. Insam J. 10 (2023)","DOI":"10.51191\/issn.2637-1898.2023.6.10.71"},{"key":"17_CR2","unstructured":"Arandas, L., et al.: all YIN no YANG: automating language-guided diffusion systems in search of abstraction. In: Explorations on Sound and New Media Art Conference, Book of Abstracts (2023)"},{"issue":"2","key":"17_CR3","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1109\/TPAMI.2018.2798607","volume":"41","author":"T Baltrusaitis","year":"2019","unstructured":"Baltrusaitis, T., Ahuja, C., Morency, L.P.: Multimodal machine learning: a survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 41(2), 423\u2013443 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Blattmann, A., et al., Align your Latents: High-Resolution Video Synthesis with Latent Diffusion Models. arXiv preprint arXiv:2304.08818, 2023","DOI":"10.1109\/CVPR52729.2023.02161"},{"issue":"2","key":"17_CR5","first-page":"1","volume":"1","author":"L Candy","year":"2006","unstructured":"Candy, L.: Practice based research: a guide. CCS Rep. 1(2), 1\u201319 (2006)","journal-title":"CCS Rep."},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"Crawford, K. and T. Paglen, Excavating AI: the politics of images in machine learning training sets. Ai & Society, 2021","DOI":"10.1007\/s00146-021-01162-8"},{"key":"17_CR7","unstructured":"Croitoru, F.-A., et al.: Diffusion models in vision: a survey. arXiv preprint arXiv:2209.04747 (2022)"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Crowson, K., et al.: VQGAN-CLIP: open domain image generation and editing with natural language guidance (2022)","DOI":"10.1007\/978-3-031-19836-6_6"},{"key":"17_CR9","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat gans on image synthesis. Adv. Neural. Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"17_CR10","unstructured":"Elgendy, M.: Deep learning for vision systems (2020)"},{"key":"17_CR11","volume-title":"Deep learning","author":"I Goodfellow","year":"2016","unstructured":"Goodfellow, I., Bengio, Y., Courville, A.: Deep learning. MIT Press, Cambridge (2016)"},{"key":"17_CR12","unstructured":"Gu, J., et al.: A systematic survey of prompt engineering on vision-language foundation models. arXiv preprint arXiv:2307.12980 (2023)"},{"key":"17_CR13","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"47","key":"17_CR14","first-page":"1","volume":"23","author":"J Ho","year":"2022","unstructured":"Ho, J., et al.: Cascaded diffusion models for high fidelity image generation. J. Mach. Learn. Res. 23(47), 1\u201333 (2022)","journal-title":"J. Mach. Learn. Res."},{"key":"17_CR15","unstructured":"Ho, J., Salimans, T.: Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598 (2022)"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Ilg, E., et al.: Flownet 2.0: evolution of optical flow estimation with deep networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.179"},{"key":"17_CR17","doi-asserted-by":"crossref","unstructured":"Khandelwal, A., et al.: Simple but effective: Clip embeddings for embodied AI. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.01441"},{"key":"17_CR18","doi-asserted-by":"crossref","unstructured":"Kim, G., Kwon, T., Ye, J.C.: DiffusionCLIP: text-guided diffusion models for robust image manipulation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.00246"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Li, L.: The impact of artificial intelligence painting on contemporary art from disco diffusion's painting creation experiment. In: 2022 International Conference on Frontiers of Artificial Intelligence and Machine Learning (FAIML). IEEE (2022)","DOI":"10.1109\/FAIML57028.2022.00020"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Li, Y., Zhang, R., Liu, J.: An enhanced prompt-based LLM reasoning scheme via knowledge graph-integrated collaboration. In: International Conference on Artificial Neural Networks. Springer (2024)","DOI":"10.1007\/978-3-031-72344-5_17"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: Adaattn: revisit attention mechanism in arbitrary neural style transfer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Liu, V., Chilton, L.B.: Design guidelines for prompt engineering text-to-image generative models. In: CHI Conference on Human Factors in Computing Systems, pp. 1\u201323 (2022)","DOI":"10.1145\/3491102.3501825"},{"key":"17_CR23","unstructured":"Liu, Y., Zhang, H., Gao, D.: DiffYOLO: object detection for anti-noise via YOLO and diffusion models. arXiv preprint arXiv:2401.01659 (2024)"},{"issue":"1","key":"17_CR24","doi-asserted-by":"publisher","first-page":"137","DOI":"10.5194\/isprs-archives-XLVIII-1-W3-2023-137-2023","volume":"48","author":"N Padkan","year":"2023","unstructured":"Padkan, N., et al.: Evaluating monocular depth estimation methods. Int. Arch. Photogrammetry, Remote Sens. Spat. Inf. Sci. 48(1), 137\u2013144 (2023)","journal-title":"Int. Arch. Photogrammetry, Remote Sens. Spat. Inf. Sci."},{"key":"17_CR25","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning. PMLR (2021)"},{"key":"17_CR26","unstructured":"Ramesh, A., et al.: Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022)"},{"key":"17_CR27","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, October 5\u20139, 2015, Proceedings, Part III 18. Springer (2015)"},{"key":"17_CR28","doi-asserted-by":"crossref","unstructured":"Rosinol, A., Leonard, J.J., Carlone, L.: Nerf-slam: real-time dense monocular slam with neural radiance fields. In: 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE (2023)","DOI":"10.1109\/IROS55552.2023.10341922"},{"key":"17_CR29","doi-asserted-by":"crossref","unstructured":"Saharia, C., et al.: Palette: image-to-image diffusion models. In: Special Interest Group on Computer Graphics and Interactive Techniques Conference Proceedings, pp. 1\u201310 (2022)","DOI":"10.1145\/3528233.3530757"},{"key":"17_CR30","unstructured":"Sohl-Dickstein, J., et al.: Deep unsupervised learning using nonequilibrium thermodynamics. In: International Conference on Machine Learning. PMLR (2015)"},{"issue":"16","key":"17_CR31","doi-asserted-by":"publisher","first-page":"3182","DOI":"10.3390\/electronics13163182","volume":"13","author":"P-C Su","year":"2024","unstructured":"Su, P.-C., Yang, M.-T.: Integrating depth-based and deep learning techniques for real-time video matting without green screens. Electronics 13(16), 3182 (2024)","journal-title":"Electronics"},{"key":"17_CR32","doi-asserted-by":"crossref","unstructured":"Wu, K., et al.: Tinyclip: clip distillation via affinity mimicking and weight inheritance. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2023)","DOI":"10.1109\/ICCV51070.2023.02008"},{"key":"17_CR33","unstructured":"Yang, L., et al.: Diffusion models: a comprehensive survey of methods and applications. arXiv preprint arXiv:2209.00796 (2022)"},{"key":"17_CR34","unstructured":"Zhang, C., et al.: Text-to-image diffusion model in generative AI: a survey. arXiv preprint arXiv:2303.07909 (2023)"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Music, Sound, Art and Design"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-90167-6_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T02:15:26Z","timestamp":1745288126000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-90167-6_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031901669","9783031901676"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-90167-6_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"20 April 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"EvoMUSART","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Trieste","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 April 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 April 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"evomusart2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.evostar.org\/2025\/evomusart\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}