{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,7]],"date-time":"2025-08-07T09:23:01Z","timestamp":1754558581565,"version":"3.41.0"},"publisher-location":"Cham","reference-count":40,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031915710","type":"print"},{"value":"9783031915727","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-91572-7_3","type":"book-chapter","created":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T08:14:56Z","timestamp":1747988096000},"page":"33-49","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["NeAT: Neural Artistic Tracing for\u00a0High Resolution Style Transfer"],"prefix":"10.1007","author":[{"given":"Dan","family":"Ruta","sequence":"first","affiliation":[]},{"given":"Andrew","family":"Gilbert","sequence":"additional","affiliation":[]},{"given":"John","family":"Collomosse","sequence":"additional","affiliation":[]},{"given":"Eli","family":"Shechtman","sequence":"additional","affiliation":[]},{"given":"Nicholas","family":"Kolkin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"key":"3_CR1","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Tov, O., Mokady, R., Gal, R., Bermano, A.H.: HyperStyle: StyleGAN inversion with hypernetworks for real image editing (2021)","DOI":"10.1109\/CVPR52688.2022.01796"},{"key":"3_CR2","unstructured":"An, J., Huang, S., Song, Y., Dou, D., Liu, W., Luo, J.: ArtFlow: Unbiased image style transfer via reversible neural flows. CoRR abs\/2103.16877 (2021). https:\/\/arxiv.org\/abs\/2103.16877"},{"key":"3_CR3","unstructured":"Chen, H., et al.: Artistic style transfer with internal-external learning and contrastive learning. In: Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems (2021). https:\/\/openreview.net\/forum?id=hm0i-cunzGW"},{"key":"3_CR4","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. arXiv preprint arXiv:2002.05709 (2020)"},{"key":"3_CR5","doi-asserted-by":"crossref","unstructured":"Collomosse, J., Bui, T., Wilber, M., Fang, C., Jin, H.: Sketching with style: Visual search with sketches and aesthetic context. In: Proceedings of ICCV (2017)","DOI":"10.1109\/ICCV.2017.290"},{"key":"3_CR6","unstructured":"Deng, Y., Tang, F., Dong, W., Huang, H., Ma, C., Xu, C.: Arbitrary video style transfer via multi-channel correlation. CoRR abs\/2009.08003 (2020). https:\/\/arxiv.org\/abs\/2009.08003"},{"key":"3_CR7","doi-asserted-by":"crossref","unstructured":"Deng, Y., et al.: Stytr$$^2$$: Image style transfer with transformers (2022)","DOI":"10.1109\/CVPR52688.2022.01104"},{"key":"3_CR8","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. CoRR abs\/2010.11929 (2020). https:\/\/arxiv.org\/abs\/2010.11929"},{"key":"3_CR9","doi-asserted-by":"publisher","unstructured":"Gafni, O., Polyak, A., Ashual, O., Sheynin, S., Parikh, D., Taigman, Y.: Make-a-scene: Scene-based text-to-image generation with human priors (2022). https:\/\/doi.org\/10.48550\/ARXIV.2203.13131, https:\/\/arxiv.org\/abs\/2203.13131","DOI":"10.48550\/ARXIV.2203.13131"},{"key":"3_CR10","doi-asserted-by":"publisher","unstructured":"Gal, R., et al.: An image is worth one word: Personalizing text-to-image generation using textual inversion (2022). https:\/\/doi.org\/10.48550\/ARXIV.2208.01618, https:\/\/arxiv.org\/abs\/2208.01618","DOI":"10.48550\/ARXIV.2208.01618"},{"key":"3_CR11","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: Proceedings of CVPR, pp. 2414\u20132423 (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"3_CR12","doi-asserted-by":"publisher","unstructured":"Hertz, A., Mokady, R., Tenenbaum, J., Aberman, K., Pritch, Y., Cohen-Or, D.: Prompt-to-prompt image editing with cross attention control (2022). https:\/\/doi.org\/10.48550\/ARXIV.2208.01626, https:\/\/arxiv.org\/abs\/2208.01626","DOI":"10.48550\/ARXIV.2208.01626"},{"key":"3_CR13","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.J.: Arbitrary style transfer in real-time with adaptive instance normalization. CoRR abs\/1703.06868 (2017). http:\/\/arxiv.org\/abs\/1703.06868","DOI":"10.1109\/ICCV.2017.167"},{"key":"3_CR14","unstructured":"Karayev, S., et al.: Recognizing image style. In: Proceedings of BMVC (2014)"},{"key":"3_CR15","doi-asserted-by":"publisher","unstructured":"Kawar, B., et al.: Imagic: Text-based real image editing with diffusion models (2022). https:\/\/doi.org\/10.48550\/ARXIV.2210.09276, https:\/\/arxiv.org\/abs\/2210.09276","DOI":"10.48550\/ARXIV.2210.09276"},{"key":"3_CR16","doi-asserted-by":"publisher","unstructured":"Kolkin, N., Kucera, M., Paris, S., Sykora, D., Shechtman, E., Shakhnarovich, G.: Neural neighbor style transfer (2022). https:\/\/doi.org\/10.48550\/ARXIV.2203.13215, https:\/\/arxiv.org\/abs\/2203.13215","DOI":"10.48550\/ARXIV.2203.13215"},{"key":"3_CR17","doi-asserted-by":"crossref","unstructured":"Kolkin, N.I., Salavon, J., Shakhnarovich, G.: Style transfer by relaxed optimal transport and self-similarity. CoRR abs\/1904.12785 (2019). http:\/\/arxiv.org\/abs\/1904.12785","DOI":"10.1109\/CVPR.2019.01029"},{"key":"3_CR18","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.: Universal style transfer via feature transforms. CoRR abs\/1705.08086 (2017). http:\/\/arxiv.org\/abs\/1705.08086"},{"key":"3_CR19","unstructured":"Liu, S., et al.: AdaAttN: Revisit attention mechanism in arbitrary neural style transfer. CoRR abs\/2108.03647 (2021). https:\/\/arxiv.org\/abs\/2108.03647"},{"key":"3_CR20","unstructured":"Luo, X., Han, Z., Yang, L., Zhang, L.: Consistent style transfer. CoRR abs\/2201.02233 (2022). https:\/\/arxiv.org\/abs\/2201.02233"},{"key":"3_CR21","doi-asserted-by":"publisher","unstructured":"Meng, C., et al.: SDEdit: Guided image synthesis and editing with stochastic differential equations (2021). https:\/\/doi.org\/10.48550\/ARXIV.2108.01073, https:\/\/arxiv.org\/abs\/2108.01073","DOI":"10.48550\/ARXIV.2108.01073"},{"key":"3_CR22","unstructured":"Park, D.Y., Lee, K.H.: Arbitrary style transfer with style-attentional networks. CoRR abs\/1812.02342 (2018). http:\/\/arxiv.org\/abs\/1812.02342"},{"key":"3_CR23","unstructured":"Park, T., et al.: Swapping autoencoder for deep image manipulation. In: Advances in Neural Information Processing Systems (2020)"},{"key":"3_CR24","doi-asserted-by":"publisher","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M.: Hierarchical text-conditional image generation with clip latents (2022). https:\/\/doi.org\/10.48550\/ARXIV.2204.06125, https:\/\/arxiv.org\/abs\/2204.06125","DOI":"10.48550\/ARXIV.2204.06125"},{"key":"3_CR25","doi-asserted-by":"crossref","unstructured":"Revaud, J., Almaz\u00e1n, J., de\u00a0Rezende, R.S., de\u00a0Souza, C.R.: Learning with average precision: Training image retrieval with a listwise loss. In: 2019 IEEE\/CVF ICCV, pp. 5106\u20135115 (2019)","DOI":"10.1109\/ICCV.2019.00521"},{"key":"3_CR26","doi-asserted-by":"publisher","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models (2021). https:\/\/doi.org\/10.48550\/ARXIV.2112.10752, https:\/\/arxiv.org\/abs\/2112.10752","DOI":"10.48550\/ARXIV.2112.10752"},{"key":"3_CR27","doi-asserted-by":"publisher","unstructured":"Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., Aberman, K.: DreamBooth: Fine tuning text-to-image diffusion models for subject-driven generation (2022). https:\/\/doi.org\/10.48550\/ARXIV.2208.12242, https:\/\/arxiv.org\/abs\/2208.12242","DOI":"10.48550\/ARXIV.2208.12242"},{"key":"3_CR28","doi-asserted-by":"crossref","unstructured":"Ruta, D., et al.: ALADIN: All layer adaptive instance normalization for fine-grained style similarity. arXiv preprint arXiv:2103.09776 (2021)","DOI":"10.1109\/ICCV48922.2021.01171"},{"key":"3_CR29","doi-asserted-by":"publisher","unstructured":"Ruta, D., et al.: StyleBabel: Artistic style tagging and captioning (2022). https:\/\/doi.org\/10.48550\/ARXIV.2203.05321, https:\/\/arxiv.org\/abs\/2203.05321","DOI":"10.48550\/ARXIV.2203.05321"},{"key":"3_CR30","doi-asserted-by":"publisher","unstructured":"Ruta, D., Gilbert, A., Motiian, S., Faieta, B., Lin, Z., Collomosse, J.: HyperNST: Hyper-networks for neural style transfer (2022). https:\/\/doi.org\/10.48550\/ARXIV.2208.04807, https:\/\/arxiv.org\/abs\/2208.04807","DOI":"10.48550\/ARXIV.2208.04807"},{"key":"3_CR31","unstructured":"Schuhmann, C., et al.: LAION-5B: An open large-scale dataset for training next generation image-text models (2022)"},{"key":"3_CR32","unstructured":"Shaham, T.R., Dekel, T., Michaeli, T.: SinGAN: learning a generative model from a single natural image. CoRR abs\/1905.01164 (2019). http:\/\/arxiv.org\/abs\/1905.01164"},{"key":"3_CR33","doi-asserted-by":"crossref","unstructured":"Sheng, L., Lin, Z., Shao, J., Wang, X.: Avatar-Net: multi-scale zero-shot style transfer by feature decoration. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01\u20139 (2018)","DOI":"10.1109\/CVPR.2018.00860"},{"key":"3_CR34","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition (2014)"},{"key":"3_CR35","doi-asserted-by":"publisher","unstructured":"Valevski, D., Kalman, M., Matias, Y., Leviathan, Y.: UniTune: Text-driven image editing by fine tuning an image generation model on a single image (2022). https:\/\/doi.org\/10.48550\/ARXIV.2210.09477, https:\/\/arxiv.org\/abs\/2210.09477","DOI":"10.48550\/ARXIV.2210.09477"},{"key":"3_CR36","unstructured":"Zhang, C., Yang, J., Wang, L., Dai, Z.: S2WAT: Image style transfer via hierarchical vision transformer using strips window attention (2022)"},{"key":"3_CR37","doi-asserted-by":"publisher","unstructured":"Zhang, H., Goodfellow, I., Metaxas, D., Odena, A.: Self-attention generative adversarial networks (2018). https:\/\/doi.org\/10.48550\/ARXIV.1805.08318, https:\/\/arxiv.org\/abs\/1805.08318","DOI":"10.48550\/ARXIV.1805.08318"},{"key":"3_CR38","doi-asserted-by":"publisher","unstructured":"Zhang, K., et al.: ARF: Artistic radiance fields (2022). https:\/\/doi.org\/10.48550\/ARXIV.2206.06360, https:\/\/arxiv.org\/abs\/2206.06360","DOI":"10.48550\/ARXIV.2206.06360"},{"key":"3_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"3_CR40","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Domain enhanced arbitrary image style transfer via contrastive learning. In: ACM SIGGRAPH (2022)","DOI":"10.1145\/3528233.3530736"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-91572-7_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T08:15:38Z","timestamp":1747988138000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-91572-7_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031915710","9783031915727"],"references-count":40,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-91572-7_3","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}