{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T23:55:40Z","timestamp":1743033340008,"version":"3.40.3"},"publisher-location":"Cham","reference-count":42,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031726637"},{"type":"electronic","value":"9783031726644"}],"license":[{"start":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T00:00:00Z","timestamp":1729900800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T00:00:00Z","timestamp":1729900800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72664-4_25","type":"book-chapter","created":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T17:02:04Z","timestamp":1729875724000},"page":"441-456","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Rejection Sampling IMLE: Designing Priors for\u00a0Better Few-Shot Image Synthesis"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3270-1533","authenticated-orcid":false,"given":"Chirag","family":"Vashist","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-8404-6392","authenticated-orcid":false,"given":"Shichong","family":"Peng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3229-271X","authenticated-orcid":false,"given":"Ke","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,26]]},"reference":[{"key":"25_CR1","unstructured":"Aghabozorgi, M., Peng, S., Li, K.: Adaptive IMLE for few-shot pretraining-free generative modelling. In: Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., Scarlett, J. (eds.) Proceedings of the 40th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol.\u00a0202, pp. 248\u2013264. PMLR, 23\u201329 July 2023. https:\/\/proceedings.mlr.press\/v202\/aghabozorgi23a.html"},{"key":"25_CR2","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. ArXiv abs\/1809.11096 (2019)"},{"key":"25_CR3","unstructured":"Child, R.: Very deep VAEs generalize autoregressive models and can outperform them on images. ArXiv abs\/2011.10650 (2021)"},{"key":"25_CR4","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat GANs on image synthesis. ArXiv abs\/2105.05233 (2021)"},{"key":"25_CR5","unstructured":"Dinh, L., Sohl-Dickstein, J.N., Bengio, S.: Density estimation using real NVP. ArXiv abs\/1605.08803 (2017)"},{"key":"25_CR6","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., Ommer, B.: Taming transformers for high-resolution image synthesis. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12868\u201312878 (2020). https:\/\/api.semanticscholar.org\/CorpusID:229297973","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"25_CR7","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27 (2014)"},{"key":"25_CR8","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: NIPS (2017)"},{"key":"25_CR9","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. ArXiv abs\/2006.11239 (2020)"},{"key":"25_CR10","unstructured":"Karras, T., et al.: Alias-free generative adversarial networks. In: NeurIPS (2021)"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"25_CR13","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of styleGAN. 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8107\u20138116 (2019). https:\/\/api.semanticscholar.org\/CorpusID:209202273","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"25_CR14","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of styleGAN. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8107\u20138116 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"25_CR15","unstructured":"Kingma, D.P., Dhariwal, P.: Glow: generative flow with invertible 1x1 convolutions. ArXiv abs\/1807.03039 (2018)"},{"key":"25_CR16","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)"},{"key":"25_CR17","doi-asserted-by":"publisher","first-page":"3964","DOI":"10.1109\/TPAMI.2020.2992934","volume":"43","author":"I Kobyzev","year":"2021","unstructured":"Kobyzev, I., Prince, S., Brubaker, M.A.: Normalizing flows: an introduction and review of current methods. IEEE Trans. Pattern Anal. Mach. Intell. 43, 3964\u20133979 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Kong, C., Kim, J., Han, D., Kwak, N.: Few-shot image generation with mixup-based distance learning (2022)","DOI":"10.1007\/978-3-031-19784-0_33"},{"key":"25_CR19","unstructured":"Kynk\u00e4\u00e4nniemi, T., Karras, T., Laine, S., Lehtinen, J., Aila, T.: Improved precision and recall metric for assessing generative models. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"25_CR20","unstructured":"Li, K., Malik, J.: Fast k-nearest neighbour search via prioritized DCI (2017)"},{"key":"25_CR21","unstructured":"Li, K., Malik, J.: Implicit maximum likelihood estimation. arXiv preprint arXiv:1809.09087 (2018)"},{"key":"25_CR22","unstructured":"Li, Y., Zhang, R., Lu, J., Shechtman, E.: Few-shot image generation with elastic weight consolidation. ArXiv abs\/2012.02780 (2020)"},{"key":"25_CR23","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"598","DOI":"10.1007\/978-3-031-19784-0_35","volume-title":"ECCV 2022","author":"Z Li","year":"2022","unstructured":"Li, Z., Wang, C., Zheng, H., Zhang, J., Li, B.: FakeCLR: exploring contrastive learning for solving latent discontinuity in data-efficient GANs. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13675, pp. 598\u2013615. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19784-0_35"},{"key":"25_CR24","unstructured":"Liu, B., Zhu, Y., Song, K., Elgammal, A.: Towards faster and stabilized GAN training for high-fidelity few-shot image synthesis. CoRR abs\/2101.04775 (2021). https:\/\/arxiv.org\/abs\/2101.04775"},{"key":"25_CR25","unstructured":"Mo, S., Cho, M., Shin, J.: Freeze discriminator: a simple baseline for fine-tuning GANs. ArXiv abs\/2002.10964 (2020)"},{"key":"25_CR26","doi-asserted-by":"crossref","unstructured":"Ojha, U., et al.: Few-shot image generation via cross-domain correspondence. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10738\u201310747 (2021)","DOI":"10.1109\/CVPR46437.2021.01060"},{"key":"25_CR27","unstructured":"van\u00a0den Oord, A., Kalchbrenner, N., Espeholt, L., Kavukcuoglu, K., Vinyals, O., Graves, A.: Conditional image generation with pixelCNN decoders. ArXiv abs\/1606.05328 (2016)"},{"key":"25_CR28","unstructured":"van\u00a0den Oord, A., Kalchbrenner, N., Kavukcuoglu, K.: Pixel recurrent neural networks. ArXiv abs\/1601.06759 (2016)"},{"key":"25_CR29","unstructured":"Razavi, A., van\u00a0den Oord, A., Vinyals, O.: Generating diverse high-fidelity images with VQ-VAE-2. ArXiv abs\/1906.00446 (2019)"},{"key":"25_CR30","unstructured":"Salimans, T., Karpathy, A., Chen, X., Kingma, D.P.: PixelCNN++: improving the pixelCNN with discretized logistic mixture likelihood and other modifications. ArXiv abs\/1701.05517 (2017)"},{"key":"25_CR31","unstructured":"Sauer, A., Chitta, K., Muller, J., Geiger, A.: Projected GANs converge faster. In: Neural Information Processing Systems (2021). https:\/\/api.semanticscholar.org\/CorpusID:240354401"},{"key":"25_CR32","doi-asserted-by":"crossref","unstructured":"Saxena, D., Cao, J., Xu, J., Kulshrestha, T.: Re-GAN: data-efficient GANs training via architectural reconfiguration. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16230\u201316240, June 2023","DOI":"10.1109\/CVPR52729.2023.01557"},{"key":"25_CR33","doi-asserted-by":"publisher","first-page":"1354","DOI":"10.1109\/TPAMI.2011.227","volume":"34","author":"Z Si","year":"2012","unstructured":"Si, Z., Zhu, S.C.: Learning hybrid image templates (HIT) by information projection. IEEE Trans. Pattern Anal. Mach. Intell. 34, 1354\u20131367 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR34","unstructured":"Song, Y., Ermon, S.: Generative modeling by estimating gradients of the data distribution. ArXiv abs\/1907.05600 (2019)"},{"key":"25_CR35","unstructured":"Song, Y., Sohl-Dickstein, J.N., Kingma, D.P., Kumar, A., Ermon, S., Poole, B.: Score-based generative modeling through stochastic differential equations. ArXiv abs\/2011.13456 (2021)"},{"key":"25_CR36","unstructured":"Vahdat, A., Kautz, J.: NVAE: A deep hierarchical variational autoencoder. ArXiv abs\/2007.03898 (2020)"},{"key":"25_CR37","doi-asserted-by":"crossref","unstructured":"Wang, Y., Gonzalez-Garcia, A., Berga, D., Herranz, L., Khan, F.S., van\u00a0de Weijer, J.: MineGAN: effective knowledge transfer from GANs to target domains with few images. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9329\u20139338 (2020)","DOI":"10.1109\/CVPR42600.2020.00935"},{"key":"25_CR38","unstructured":"Yang, M., Wang, Z., Chi, Z., Zhang, Y.: FreGAN: exploiting frequency components for training GANs under limited data (2022)"},{"key":"25_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"25_CR40","unstructured":"Zhao, M., Cong, Y., Carin, L.: On leveraging pretrained GANs for generation with limited data. In: ICML (2020)"},{"key":"25_CR41","unstructured":"Zhao, S., Liu, Z., Lin, J., Zhu, J.Y., Han, S.: Differentiable augmentation for data-efficient GAN training. ArXiv abs\/2006.10738 (2020)"},{"key":"25_CR42","unstructured":"Zhu, P., Abdal, R., Qin, Y., Wonka, P.: Improved styleGAN embedding: where are the good latents? ArXiv abs\/2012.09036 (2020)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72664-4_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T17:08:59Z","timestamp":1729876139000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72664-4_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,26]]},"ISBN":["9783031726637","9783031726644"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72664-4_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,10,26]]},"assertion":[{"value":"26 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}