{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:27:09Z","timestamp":1775579229453,"version":"3.50.1"},"reference-count":215,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T00:00:00Z","timestamp":1717113600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T00:00:00Z","timestamp":1717113600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52071349"],"award-info":[{"award-number":["52071349"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52071349"],"award-info":[{"award-number":["52071349"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52071349"],"award-info":[{"award-number":["52071349"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52071349"],"award-info":[{"award-number":["52071349"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52071349"],"award-info":[{"award-number":["52071349"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Independent research projects for postgraduate students","award":["SZKY2023054"],"award-info":[{"award-number":["SZKY2023054"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Artif Intell Rev"],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Diffusion models (DMs) are a type of potential generative models, which have achieved better effects in many fields than traditional methods. DMs consist of two main processes: one is the forward process of gradually adding noise to the original data until pure Gaussian noise; the other is the reverse process of gradually removing noise to generate samples conforming to the target distribution. DMs optimize the application results through the iterative noise processing process. However, this greatly increases the computational and storage costs in the training and inference stages, limiting the wide application of DMs. Therefore, how to effectively reduce the resource consumption of using DMs while giving full play to their good performance has become a valuable and necessary research problem. At present, some research has been devoted to lightweight DMs to solve this problem, but there has been no survey in this area. This paper focuses on lightweight DMs methods in the field of image processing, classifies them according to their processing ideas. Finally, the development prospect of future work is analyzed and discussed. It is hoped that this paper can provide other researchers with strategic ideas to reduce the resource consumption of DMs, thereby promoting the further development of this research direction and providing available models for wider applications.<\/jats:p>","DOI":"10.1007\/s10462-024-10800-8","type":"journal-article","created":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T05:01:36Z","timestamp":1717131696000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":22,"title":["Lightweight diffusion models: a survey"],"prefix":"10.1007","volume":"57","author":[{"given":"Wei","family":"Song","sequence":"first","affiliation":[]},{"given":"Wen","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yanghao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xiaobing","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,31]]},"reference":[{"key":"10800_CR1","unstructured":"Ambrogioni L (2023) The statistical thermodynamics of generative diffusion models. Preprint. arXiv:2310.17467"},{"key":"10800_CR2","unstructured":"Arakawa S, Tsunashima H, Horita D, Tanaka K, Morishima S (2023) Memory efficient diffusion probabilistic models via patch-based generation. Preprint. arXiv:2304.07087"},{"key":"10800_CR4","doi-asserted-by":"crossref","unstructured":"Avrahami O, Lischinski D, Fried O (2022) Blended diffusion for text-driven editing of natural images. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18208\u201318218","DOI":"10.1109\/CVPR52688.2022.01767"},{"issue":"4","key":"10800_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3592450","volume":"42","author":"O Avrahami","year":"2023","unstructured":"Avrahami O, Fried O, Lischinski D (2023) Blended latent diffusion. ACM Trans Graph (TOG) 42(4):1\u201311","journal-title":"ACM Trans Graph (TOG)"},{"key":"10800_CR5","unstructured":"Bai H, Hou L, Shang L, Jiang X, King I, Lyu MR (2022) Towards efficient post-training quantization of pre-trained language models. In: Advances in neural information processing systems, vol 35, pp 1405\u20131418"},{"key":"10800_CR6","unstructured":"Balaji Y, Nah S, Huang X, Vahdat A, Song J, Kreis K, Aittala M, Aila T, Laine S, Catanzaro B et al (2022) ediffi: text-to-image diffusion models with an ensemble of expert denoisers. Preprint. arXiv:2211.01324"},{"key":"10800_CR7","unstructured":"Bao F, Li C, Sun J, Zhu J, Zhang B (2022) Estimating the optimal covariance with imperfect mean in diffusion probabilistic models. In: International conference on machine learning, pp 1555\u20131584. PMLR"},{"key":"10800_CR8","doi-asserted-by":"crossref","unstructured":"Bao F, Nie S, Xue K, Cao Y, Li C, Su H, Zhu J (2023) All are worth words: a vit backbone for diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 22669\u201322679","DOI":"10.1109\/CVPR52729.2023.02171"},{"key":"10800_CR9","unstructured":"Baykal G, Karagoz H.F, Binhuraib T, Unal G (2023) ProtoDiffusion: classifier-free diffusion guidance with prototype learning. Preprint. arXiv:2307.01924"},{"key":"10800_CR10","unstructured":"Berthelot D, Autef A, Lin J, Yap DA, Zhai S, Hu S, Zheng D, Talbot W, Gu E (2023) TRACT: denoising diffusion models with transitive closure time-distillation. Preprint. arXiv:2303.04248"},{"key":"10800_CR11","doi-asserted-by":"crossref","unstructured":"Blattmann A, Rombach R, Ling H, Dockhorn T, Kim SW, Fidler S, Kreis K (2023) Align your latents: high-resolution video synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 22563\u201322575","DOI":"10.1109\/CVPR52729.2023.02161"},{"key":"10800_CR12","doi-asserted-by":"crossref","unstructured":"Bolya D, Hoffman J (2023) Token merging for fast stable diffusion. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4598\u20134602","DOI":"10.1109\/CVPRW59228.2023.00484"},{"key":"10800_CR13","doi-asserted-by":"crossref","unstructured":"Bond-Taylor S, Hessey P, Sasaki H, Breckon TP, Willcocks CG (2022) Unleashing transformers: parallel token prediction with discrete absorbing diffusion for fast high-resolution image generation from vector-quantized codes. In: European conference on computer vision. Springer, Cham, pp 170\u2013188","DOI":"10.1007\/978-3-031-20050-2_11"},{"key":"10800_CR14","unstructured":"Chang H, Shen H, Cai Y, Ye X, Xu Z, Cheng W, Lv K, Zhang W, Lu Y, Guo H (2023) Effective Quantization for Diffusion Models on CPUs. Preprint. arXiv:2311.16133"},{"issue":"11","key":"10800_CR15","doi-asserted-by":"publisher","first-page":"1021","DOI":"10.3844\/jcssp.2022.1021.1029","volume":"18","author":"EH Cheikh Tourad","year":"2022","unstructured":"Cheikh Tourad EH (2022) Eleuldj M Quantization and pipelined hardware implementation of deep neural network models. J Comput Sci 18(11):1021\u20131029. https:\/\/doi.org\/10.3844\/jcssp.2022.1021.1029","journal-title":"J Comput Sci"},{"key":"10800_CR18","unstructured":"Chen N, Zhang Y, Zen H, Weiss R.J, Norouzi M, Chan W (2020) WaveGrad: estimating gradients for waveform generation. In: International Conference on Learning Representations"},{"key":"10800_CR16","doi-asserted-by":"crossref","unstructured":"Chen D, Mei J-P, Zhang H, Wang C, Feng Y, Chen C (2022a) Knowledge distillation with the reused teacher classifier. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11933\u201311942","DOI":"10.1109\/CVPR52688.2022.01163"},{"key":"10800_CR17","unstructured":"Chen Z, Wu Y, Leng Y, Chen J, Liu H, Tan X, Cui Y, Wang K, He L, Zhao S et al (2022b) Resgrad: residual denoising diffusion probabilistic models for text to speech. Preprint. arXiv:2212.14518"},{"key":"10800_CR19","unstructured":"Chin T-W, Ding R, Zhang C, Marculescu D (2020) Towards efficient model compression via learned global ranking. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1518\u20131528"},{"issue":"6","key":"10800_CR20","doi-asserted-by":"publisher","first-page":"763","DOI":"10.1017\/S1351324921000322","volume":"27","author":"KW Church","year":"2021","unstructured":"Church KW, Chen Z, Ma Y (2021) Emerging trends: a gentle introduction to fine-tuning. Nat Lang Eng 27(6):763\u2013778","journal-title":"Nat Lang Eng"},{"key":"10800_CR21","unstructured":"Clark K, Jaini P (2023) Text-to-image diffusion models are zero-shot classifiers. In: ICLR 2023 workshop on mathematical and empirical understanding of foundation models"},{"key":"10800_CR22","unstructured":"Corso G, Xu Y, De\u00a0Bortoli V, Barzilay R, Jaakkola T (2023) Particle guidance: non-IID diverse sampling with diffusion models. In: NeurIPS 2023 workshop on deep learning and inverse problems"},{"key":"10800_CR23","unstructured":"Dhariwal P, Nichol A (2021) Diffusion models beat gans on image synthesis. In: Advances in neural information processing systems, vol 34, pp 8780\u20138794"},{"key":"10800_CR24","unstructured":"Dockhorn T, Vahdat A (2022) Genie: higher-order denoising diffusion solvers. In: Advances in neural information processing systems, vol 35, pp 30150\u201330166"},{"key":"10800_CR25","unstructured":"Dockhorn T, Vahdat A, Kreis K (2021) Score-based generative modeling with critically-damped langevin diffusion. In: International conference on learning representations"},{"key":"10800_CR26","doi-asserted-by":"crossref","unstructured":"Esser P, Rombach R, Ommer B (2021) Taming transformers for high-resolution image synthesis. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12873\u201312883","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"10800_CR30","unstructured":"Fan Y, Lee K (2023) Optimizing ddpm sampling with shortcut fine-tuning. Preprint. arXiv:2301.13362"},{"key":"10800_CR27","unstructured":"Fan W-C, Chen Y-C, Chen D, Cheng Y, Yuan L, Wang Y-CF (2023) Frido: feature pyramid diffusion for complex scene image synthesis. In: Proceedings of the AAAI conference on artificial intelligence, vol 37, pp 579\u2013587"},{"key":"10800_CR31","unstructured":"Fan Y, Watkins O, Du Y, Liu H, Ryu M, Boutilier C, Abbeel P, Ghavamzadeh M, Lee K, Lee K (2023) DPOK: reinforcement learning for fine-tuning text-to-image diffusion models. Preprint. arXiv:2305.16381"},{"key":"10800_CR28","doi-asserted-by":"crossref","unstructured":"Fang G, Ma X, Song M, Mi MB, Wang X (2023a) DepGraph: towards any structural pruning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, p 16091\u201316101","DOI":"10.1109\/CVPR52729.2023.01544"},{"key":"10800_CR29","unstructured":"Fang G, Ma X, Wang X (2023b) Structural pruning for diffusion models. Preprint. arXiv:2305.10924"},{"key":"10800_CR32","doi-asserted-by":"crossref","unstructured":"Gao Q, Li Z, Zhang J, Zhang Y, Shan H (2023a) CoreDiff: contextual error-modulated generalized diffusion model for low-dose CT denoising and generalization. Preprint. arXiv:2304.01814","DOI":"10.1109\/TMI.2023.3320812"},{"key":"10800_CR33","unstructured":"Gao Y, Pan Z, Zhou X, Kang L, Chaudhari P (2023b) Fast diffusion probabilistic model sampling through the lens of backward error analysis. Preprint. arXiv:2304.11446"},{"key":"10800_CR34","unstructured":"Ghimire S, Liu J, Comas A, Hill D, Masoomi A, Camps O, Dy J (2023) Geometry of score based generative models. Preprint. arXiv:2302.04411"},{"key":"10800_CR35","doi-asserted-by":"crossref","unstructured":"Go H, Lee Y, Kim J-Y, Lee S, Jeong M, Lee HS, Choi S (2023) Towards practical plug-and-play diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1962\u20131971","DOI":"10.1109\/CVPR52729.2023.00195"},{"key":"10800_CR36","unstructured":"Golnari PA, Yao Z, He Y (2023) Selective guidance: are all the denoising steps of guided diffusion important? Preprint. arXiv:2305.09847"},{"key":"10800_CR37","unstructured":"Gong S, Li M, Feng J, Wu Z, Kong L (2022) Diffuseq: sequence to sequence text generation with diffusion models. In: The 11th International conference on learning representations"},{"key":"10800_CR38","unstructured":"Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Proceedings of the 27th international conference on neural information processing systems, vol 2, pp 2672\u20132680"},{"key":"10800_CR39","doi-asserted-by":"crossref","unstructured":"Gu S, Chen D, Bao J, Wen F, Zhang B, Chen D, Yuan L, Guo B (2022) Vector quantized diffusion model for text-to-image synthesis. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10696\u201310706","DOI":"10.1109\/CVPR52688.2022.01043"},{"issue":"4","key":"10800_CR40","doi-asserted-by":"publisher","first-page":"3313","DOI":"10.1109\/TKDE.2021.3130191","volume":"35","author":"J Gui","year":"2023","unstructured":"Gui J, Sun Z, Wen Y, Tao D, Ye J (2023) A review on generative adversarial networks: algorithms, theory, and applications. IEEE Trans Knowl Data Eng 35(4):3313\u20133332","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"10800_CR41","unstructured":"Guo HA, Lu C, Bao F, Pang T, Shuicheng Y, Du C, Li C (2023) Gaussian mixture solvers for diffusion models. In: 37th Conference on neural information processing systems"},{"key":"10800_CR42","unstructured":"Guth F, Coste S, De Bortoli V, Mallat S (2022) Wavelet score-based generative modeling. In: Advances in neural information processing systems, vol 35, pp 478\u2013491"},{"key":"10800_CR44","doi-asserted-by":"crossref","unstructured":"Han L, Li Y, Zhang H, Milanfar P, Metaxas D, Yang F (2023) SVDiff: compact parameter space for diffusion fine-tuning. Preprint. arXiv:2303.11305","DOI":"10.1109\/ICCV51070.2023.00673"},{"key":"10800_CR43","doi-asserted-by":"crossref","unstructured":"Hang T, Gu S, Li C, Bao J, Chen D, Hu H, Geng X, Guo B (2023) Efficient diffusion training via Min-SNR weighting strategy. Preprint. arXiv:2303.09556","DOI":"10.1109\/ICCV51070.2023.00684"},{"key":"10800_CR45","unstructured":"Haxholli E, Lorenzi M (2023) Faster training of diffusion models and improved density estimation via parallel score matching. Preprint. arXiv:2306.02658"},{"key":"10800_CR46","unstructured":"He Y, Liu L, Liu J, Wu W, Zhou H, Zhuang B (2023a) PTQD: accurate post-training quantization for diffusion models. Preprint. arXiv:2305.10657"},{"key":"10800_CR47","unstructured":"He Y, Liu J, Wu W, Zhou H, Zhuang B (2023b) EfficientDM: efficient quantization-aware fine-tuning of low-bit diffusion models. Preprint. arXiv:2310.03270"},{"key":"10800_CR48","doi-asserted-by":"crossref","unstructured":"He J, Liu J, Ye Z, Huang R, Cui C, Liu H, Zhao Z (2023c) RMSSinger: realistic-music-score based singing voice synthesis. Preprint. arXiv:2305.10686","DOI":"10.18653\/v1\/2023.findings-acl.16"},{"key":"10800_CR50","unstructured":"He L, Yan H, Luo M, Luo K, Wang W, Du W, Chen H, Yang H, Zhang Y (2023d) Iterative reconstruction based on latent diffusion model for sparse data reconstruction. Preprint. arXiv:2307.12070"},{"key":"10800_CR49","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local Nash equilibrium. In: Advances in neural information processing systems, vol 30, pp 6626\u20136637"},{"key":"10800_CR54","unstructured":"Ho J, Salimans T (2021) Classifier-free diffusion guidance. In: NeurIPS 2021 Workshop on Deep Generative Models and Downstream Applications"},{"key":"10800_CR51","unstructured":"Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. In: Advances in neural information processing systems, vol 33, pp 6840\u20136851"},{"issue":"1","key":"10800_CR52","first-page":"2249","volume":"23","author":"J Ho","year":"2022","unstructured":"Ho J, Saharia C, Chan W, Fleet DJ, Norouzi M (2022) Salimans T Cascaded diffusion models for high fidelity image generation. J Mach Learn Res 23(1):2249\u20132281","journal-title":"J Mach Learn Res"},{"key":"10800_CR53","unstructured":"Hoogeboom E, Heek J, Salimans T (2023) simple diffusion: end-to-end diffusion for high resolution images. Preprint. arXiv:2301.11093"},{"key":"10800_CR58","unstructured":"Hu E.J, Wallis P, Allen-Zhu Z, Li Y, Wang S, Wang L, Chen W et al (2021) Lora: low-rank adaptation of large language models. In: International conference on learning representations"},{"key":"10800_CR55","unstructured":"Huang T, You S, Wang F, Qian C, Xu C (2022) Knowledge distillation from a stronger teacher. In: Advances in neural information processing systems, vol 35, pp 33716\u201333727"},{"key":"10800_CR56","unstructured":"Huang Y, Qin Z, Liu X, Xu K (2023a) Decoupled diffusion models with explicit transition probability. Preprint. arXiv:2306.13720"},{"key":"10800_CR57","unstructured":"Huang Z, Zhou P, Shuicheng Y, Lin L (2023b) Scalelong: towards more stable training of diffusion model via scaling network long skip connection. In: 37th Conference on neural information processing systems"},{"key":"10800_CR59","unstructured":"Jeon J, Park N (2022) SPI-GAN: distilling score-based generative models with straight-path interpolations. Preprint. arXiv:2206.14464"},{"key":"10800_CR62","unstructured":"Jia X, Zhao Y, Chan KC, Li Y, Zhang H, Gong B, Hou T, Wang H, Su Y-C (2023) Taming encoder for zero fine-tuning image customization with text-to-image diffusion models. Preprint. arXiv:2304.02642"},{"key":"10800_CR60","unstructured":"Jiang C, Hui B, Liu B, Yan D (2023a) Successfully applying lottery ticket hypothesis to diffusion model. Preprint. arXiv:2310.18823"},{"key":"10800_CR61","doi-asserted-by":"crossref","unstructured":"Jiang H, Luo A, Han S, Fan H, Liu S (2023b) Low-light image enhancement with wavelet-based diffusion models. Preprint. arXiv:2306.00306","DOI":"10.1145\/3618373"},{"key":"10800_CR63","unstructured":"Jolicoeur-Martineau A, Li K, Pich\u00e9-Taillefer R, Kachman T, Mitliagkas I (2021) Gotta go fast when generating data with score-based models. Preprint. arXiv:2105.14080"},{"key":"10800_CR64","unstructured":"Kang J, Choi J, Choi S, Han B (2023) Observation-guided diffusion probabilistic models. Preprint. arXiv:2310.04041"},{"key":"10800_CR65","unstructured":"Karras T, Aittala M, Aila T, Laine S (2022) Elucidating the design space of diffusion-based generative models. In: Advances in neural information processing systems, vol 35, pp 26565\u201326577"},{"key":"10800_CR66","unstructured":"Kim B-K, Song H-K, Castells T, Choi S (2023) On architectural compression of text-to-image diffusion models. Preprint. arXiv:2305.15798"},{"key":"10800_CR67","unstructured":"Kim D, Na B, Kwon SJ, Lee D, Kang W, Moon I (2022a) Maximum likelihood training of implicit nonlinear diffusion model. In: Advances in neural information processing systems, vol 35, pp 32270\u201332284"},{"key":"10800_CR68","doi-asserted-by":"crossref","unstructured":"Kim G, Kwon T, Ye JC (2022b) Diffusionclip: Text-guided diffusion models for robust image manipulation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2426\u20132435","DOI":"10.1109\/CVPR52688.2022.00246"},{"key":"10800_CR69","unstructured":"Kim D, Lai C-H, Liao W-H, Murata N, Takida Y, Uesaka T, He Y, Mitsufuji Y, Ermon S (2023) Consistency trajectory models: Learning probability flow ode trajectory of diffusion. Preprint. arXiv:2310.02279"},{"key":"10800_CR70","unstructured":"Kingma DP, Dhariwal P (2018) Glow: generative flow with invertible 1 \u00d7 1 convolutions. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, 10236\u201310245"},{"key":"10800_CR71","unstructured":"Kingma DP, Welling M (2013) Auto-encoding variational Bayes. Preprint. arXiv:1312.6114"},{"key":"10800_CR72","unstructured":"Kingma D, Salimans T, Poole B, Ho J (2021) Variational diffusion models. In: Advances in neural information processing systems, vol 34, pp 21696\u201321707"},{"key":"10800_CR73","unstructured":"Kong Z, Ping W (2021) On fast sampling of diffusion probabilistic models. Preprint. arXiv:2106.00132"},{"key":"10800_CR74","unstructured":"Krizhevsky A, Hinton G et al (2009) Learning multiple layers of features from tiny images. Technical Report"},{"key":"10800_CR75","unstructured":"Kulikov V, Yadin S, Kleiner M, Michaeli T (2023) SinDDM: a single image denoising diffusion model. In: International conference on machine learning, pp 17920\u201317930. PMLR"},{"key":"10800_CR77","unstructured":"Kumar A, Raghunathan A, Jones R, Ma T, Liang P (2022) Fine-tuning can distort pretrained features and underperform out-of-distribution. In: International conference on learning representations"},{"key":"10800_CR76","doi-asserted-by":"crossref","unstructured":"Kumari N, Zhang B, Zhang R, Shechtman E, Zhu J-Y (2023) Multi-concept customization of text-to-image diffusion. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1931\u20131941","DOI":"10.1109\/CVPR52729.2023.00192"},{"key":"10800_CR78","unstructured":"Kynk\u00e4\u00e4nniemi T, Karras T, Laine S, Lehtinen J, Aila T (2019) Improved precision and recall metric for assessing generative models. In: Advances in neural information processing systems, vol 32"},{"key":"10800_CR79","unstructured":"Laousy O, Araujo A, Chassagnon G, Revel M-P, Garg S, Khorrami F, Vakalopoulou M (2023) Towards better certified segmentation via diffusion models. Preprint. arXiv:2306.09949"},{"key":"10800_CR80","doi-asserted-by":"crossref","unstructured":"Lee Y, Kim J-Y, Go H, Jeong M, Oh S, Choi S (2023a) Multi-architecture multi-expert diffusion models. Preprint. arXiv:2306.04990","DOI":"10.1609\/aaai.v38i12.29245"},{"key":"10800_CR81","unstructured":"Lee S, Kim B, Ye JC (2023b) Minimizing trajectory curvature of ODE-based generative models. Preprint. arXiv:2301.12003"},{"key":"10800_CR82","unstructured":"Lee K, Liu H, Ryu M, Watkins O, Du Y, Boutilier C, Abbeel P, Ghavamzadeh M, Gu SS (2023c) Aligning text-to-image models using human feedback. Preprint. arXiv:2302.12192"},{"key":"10800_CR83","unstructured":"Lee Y, Park K, Cho Y, Lee Y-J, Hwang SJ (2023d) KOALA: self-attention matters in knowledge distillation of latent diffusion models for memory-efficient and fast image synthesis. Preprint. arXiv:2312.04005"},{"key":"10800_CR84","doi-asserted-by":"crossref","unstructured":"Lemercier J-M, Richter J, Welker S, Gerkmann T (2023) StoRM: a diffusion-based stochastic regeneration model for speech enhancement and dereverberation. IEEE\/ACM Trans Audio Speech Lang Process 31:2724\u20132737","DOI":"10.1109\/TASLP.2023.3294692"},{"key":"10800_CR85","doi-asserted-by":"crossref","unstructured":"Leng Y, Huang Q, Wang Z, Liu Y, Zhang H (2023) DiffuseGAE: controllable and high-fidelity image manipulation from disentangled representation. Preprint. arXiv:2307.05899","DOI":"10.1145\/3595916.3626402"},{"key":"10800_CR86","unstructured":"Li X, Thickstun J, Gulrajani I, Liang PS, Hashimoto TB (2022a) Diffusion-LM improves controllable text generation. In: Advances in neural information processing systems, vol 35, pp 4328\u20134343"},{"key":"10800_CR87","unstructured":"Li M, Lin J, Meng C, Ermon S, Han S, Zhu J-Y (2022b) Efficient spatially sparse inference for conditional GANs and diffusion models. In: Advances in neural information processing systems, vol 35, pp 28858\u201328873"},{"key":"10800_CR88","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1016\/j.neucom.2022.01.029","volume":"479","author":"H Li","year":"2022","unstructured":"Li H, Yang Y, Chang M, Chen S, Feng H, Xu Z, Li Q, Chen Y (2022c) SRDiff: single image super-resolution with diffusion probabilistic models. Neurocomputing 479:47\u201359","journal-title":"Neurocomputing"},{"key":"10800_CR96","unstructured":"Li W, Yu X, Zhou K, Song Y, Lin Z, Jia J (2022d) Image Inpainting via Iteratively decoupled probabilistic modeling. Preprint. arXiv:2212.02963"},{"key":"10800_CR90","doi-asserted-by":"crossref","unstructured":"Li X, Lian L, Liu Y, Yang H, Dong Z, Kang D, Zhang S, Keutzer K (2023a) Q-Diffusion: quantizing diffusion models. Preprint. arXiv:2302.04304","DOI":"10.1109\/ICCV51070.2023.01608"},{"key":"10800_CR95","unstructured":"Li Y, Wang H, Jin Q, Hu J, Chemerys P, Fu Y, Wang Y, Tulyakov S, Ren J (2023b) SnapFusion: text-to-image diffusion model on mobile devices within two seconds. Preprint. arXiv:2306.00980"},{"key":"10800_CR89","doi-asserted-by":"crossref","unstructured":"Liang J, Zeng H, Zhang L (2022) Efficient and degradation-adaptive network for real-world image super-resolution. In: European conference on computer vision. Springer, Cham, pp 574\u2013591","DOI":"10.1007\/978-3-031-19797-0_33"},{"key":"10800_CR91","doi-asserted-by":"crossref","unstructured":"Lin S, Liu B, Li J, Yang X (2024) Common diffusion noise schedules and sample steps are flawed. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 5404\u20135411","DOI":"10.1109\/WACV57701.2024.00532"},{"key":"10800_CR93","doi-asserted-by":"crossref","unstructured":"Liu X, Ji K, Fu Y, Tam W, Du Z, Yang Z, Tang J (2022) P-Tuning: prompt tuning can be comparable to fine-tuning across scales and tasks. In: Proceedings of the 60th annual meeting of the association for computational linguistics, vol 2: short papers, pp 61\u201368","DOI":"10.18653\/v1\/2022.acl-short.8"},{"key":"10800_CR92","doi-asserted-by":"crossref","unstructured":"Liu Z, Guo Y, Yu K (2023a) Diffvoice: Text-to-speech with latent diffusion. In: ICASSP 2023-2023 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, pp 1\u20135","DOI":"10.1109\/ICASSP49357.2023.10095100"},{"key":"10800_CR94","unstructured":"Liu E, Ning X, Lin Z, Yang H, Wang Y (2023b) OMS-DPM: optimizing the model schedule for diffusion probabilistic models. Preprint. arXiv:2306.08860"},{"key":"10800_CR97","unstructured":"Lu C, Zhou Y, Bao F, Chen J, Li C, Zhu J (2022a) DPM-Solver: a fast ODE solver for diffusion probabilistic model sampling in around 10 steps. In: Advances in neural information processing systems, vol 35, pp 5775\u20135787"},{"key":"10800_CR103","unstructured":"Lu C, Zhou Y, Bao F, Chen J, Li C, Zhu J (2022b) DPM-Solver++: fast solver for guided sampling of diffusion probabilistic models. Preprint. arXiv:2211.01095"},{"key":"10800_CR99","doi-asserted-by":"crossref","unstructured":"Lu S, Liu Y, Kong AW-K (2023) TF-ICON: diffusion-based training-free cross-domain image composition. Preprint. arXiv:2307.12493","DOI":"10.1109\/ICCV51070.2023.00218"},{"key":"10800_CR98","unstructured":"Luhman E, Luhman T (2021) Knowledge distillation in iterative generative models for improved sampling speed. Preprint. arXiv:2101.02388"},{"key":"10800_CR102","unstructured":"Luo F, Xiang J, Zhang J, Han X, Yang W (2023a) Image super-resolution via latent diffusion: a sampling-space mixture of experts and frequency-augmented decoder approach. Preprint. arXiv:2310.12004"},{"key":"10800_CR100","unstructured":"Luo S, Tan Y, Huang L, Li J, Zhao H (2023b) Latent consistency models: synthesizing high-resolution images with few-step inference. Preprint. arXiv:2310.04378"},{"key":"10800_CR101","unstructured":"Luo S, Tan Y, Patil S, Gu D, Platen P, Passos A, Huang L, Li J, Zhao H (2023c) LCM-LoRA: a universal stable-diffusion acceleration module. Preprint. arXiv:2311.05556"},{"key":"10800_CR106","unstructured":"Ma J, Hu T, Wang W, Sun J (2023a) Elucidating the design space of classifier-guided diffusion generation. Preprint. arXiv:2310.11311"},{"key":"10800_CR107","doi-asserted-by":"crossref","unstructured":"Ma Z, Li J, Zhou B et al (2023b) LMD: faster image reconstruction with latent masking diffusion. Preprint. arXiv:2312.07971","DOI":"10.1609\/aaai.v38i5.28209"},{"key":"10800_CR104","doi-asserted-by":"crossref","unstructured":"Macha S, Oza O, Escott A, Caliva F, Armitano R, Cheekatmalla SK, Parthasarathi SHK, Liu Y (2023) Fixed-point quantization aware training for on-device keyword-spotting. In: ICASSP 2023-2023 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, pp 1\u20135","DOI":"10.1109\/ICASSP49357.2023.10095977"},{"key":"10800_CR105","doi-asserted-by":"crossref","unstructured":"Mahajan D, Girshick R, Ramanathan V, He K, Paluri M, Li Y, Bharambe A, Van Der\u00a0Maaten L (2018) Exploring the limits of weakly supervised pretraining. In: Proceedings of the European conference on computer vision (ECCV), pp 181\u2013196","DOI":"10.1007\/978-3-030-01216-8_12"},{"key":"10800_CR108","unstructured":"Manukyan H, Sargsyan A, Atanyan B, Wang Z, Navasardyan S, Shi H (2023) HD-Painter: high-resolution and prompt-faithful text-guided image inpainting with diffusion models. Preprint. arXiv:2312.14091"},{"key":"10800_CR109","doi-asserted-by":"crossref","unstructured":"Mao W, Xu C, Zhu Q, Chen S, Wang Y (2023) Leapfrog diffusion model for stochastic trajectory prediction. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5517\u20135526","DOI":"10.1109\/CVPR52729.2023.00534"},{"key":"10800_CR110","unstructured":"Masip S, Rodriguez P, Tuytelaars T, Ven GM (2023) Continual learning of diffusion models with generative distillation. Preprint. arXiv:2311.14028"},{"key":"10800_CR111","doi-asserted-by":"crossref","unstructured":"Mei K, Patel V (2023) VIDM: video implicit diffusion models. In: Proceedings of the AAAI conference on artificial intelligence, vol 37, pp 9117\u20139125","DOI":"10.1609\/aaai.v37i8.26094"},{"key":"10800_CR112","unstructured":"Mei K, Delbracio M, Talebi H, Tu Z, Patel VM, Milanfar P (2023) Conditional diffusion distillation. Preprint. arXiv:2310.01407"},{"key":"10800_CR113","unstructured":"Meng C, He Y, Song Y, Song J, Wu J, Zhu J-Y, Ermon S (2021) SDEdit: guided image synthesis and editing with stochastic differential equations. In: International conference on learning representations"},{"key":"10800_CR114","doi-asserted-by":"crossref","unstructured":"Meng C, Rombach R, Gao R, Kingma D, Ermon S, Ho J, Salimans T (2023) On distillation of guided diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14297\u201314306","DOI":"10.1109\/CVPR52729.2023.01374"},{"key":"10800_CR115","unstructured":"Moser B.B, Frolov S, Raue F, Palacio S, Dengel A (2023) Waving goodbye to low-res: a diffusion-wavelet approach for image super-resolution. CoRR. arXiv:2304.01994"},{"key":"10800_CR116","unstructured":"Nash C, Menick J, Dieleman S, Battaglia P (2021) Generating images with sparse representations. In: International conference on machine learning, pp 7958\u20137968. PMLR"},{"key":"10800_CR117","unstructured":"Nguyen TH, Tran A (2023) SwiftBrush: one-step text-to-image diffusion model with variational score distillation. Preprint. arXiv:2312.05239"},{"key":"10800_CR122","doi-asserted-by":"crossref","unstructured":"Ni H, Shi C, Li K, Huang SX, Min MR (2023) Conditional image-to-video generation with latent flow diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18444\u201318455","DOI":"10.1109\/CVPR52729.2023.01769"},{"key":"10800_CR118","unstructured":"Nichol AQ, Dhariwal P (2021) Improved denoising diffusion probabilistic models. In: International conference on machine learning, pp 8162\u20138171. PMLR"},{"key":"10800_CR119","unstructured":"Nichol A.Q, Dhariwal P, Ramesh A, Shyam P, Mishkin P, Mcgrew B, Sutskever I, Chen M (2022) Glide: towards photorealistic image generation and editing with text-guided diffusion models. In: International conference on machine learning, pp 16784\u201316804. PMLR"},{"key":"10800_CR120","unstructured":"Nie S, Guo HA, Lu C, Zhou Y, Zheng C, Li C (2023) The blessing of randomness: SDE beats ODE in general diffusion-based image editing. Preprint. arXiv:2311.01410"},{"key":"10800_CR121","unstructured":"Ning M, Sangineto E, Porrello A, Calderara S, Cucchiara R (2023) Input perturbation reduces exposure bias in diffusion models. Preprint. arXiv:2301.11706"},{"key":"10800_CR123","unstructured":"Niu A, Trung PX, Zhang K, Sun J, Zhu Y, Kweon IS, Zhang Y (2023) ACDMSR: accelerated conditional diffusion models for single image super-resolution. Preprint. arXiv:2307.00781"},{"key":"10800_CR124","doi-asserted-by":"crossref","unstructured":"Oh S, Sim H, Kim J, Lee J (2022) Non-uniform step size quantization for accurate post-training quantization. In: European conference on computer vision. Springer, Cham, pp 658\u2013673","DOI":"10.1007\/978-3-031-20083-0_39"},{"key":"10800_CR125","doi-asserted-by":"crossref","unstructured":"Okamoto T, Toda T, Shiga Y, Kawai H (2021) Noise level limited sub-modeling for diffusion probabilistic vocoders. In: ICASSP 2021-2021 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, pp 6029\u20136033","DOI":"10.1109\/ICASSP39728.2021.9415087"},{"key":"10800_CR126","unstructured":"Pandey K, Mukherjee A, Rai P, Kumar A (2022) DiffuseVAE: efficient, controllable and high-fidelity generation from low-dimensional latents. Preprint. arXiv:2201.00308"},{"key":"10800_CR127","unstructured":"Pandey K, Rudolph M, Mandt S (2023) Efficient integrators for diffusion generative models. Preprint. arXiv:2310.07894"},{"key":"10800_CR128","unstructured":"Peebles W, Xie S (2022) Scalable diffusion models with transformers. Preprint. arXiv:2212.09748"},{"key":"10800_CR129","unstructured":"Permenter F, Yuan C (2023) Interpreting and improving diffusion models using the Euclidean distance function. Preprint. arXiv:2306.04848"},{"key":"10800_CR130","doi-asserted-by":"crossref","unstructured":"Phung H, Dao Q, Tran A (2023) Wavelet diffusion models are fast and scalable image generators. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10199\u201310208","DOI":"10.1109\/CVPR52729.2023.00983"},{"key":"10800_CR131","unstructured":"Popov V, Vovk I, Gogoryan V, Sadekova T, Kudinov M (2021) Grad-TTS: a diffusion probabilistic model for text-to-speech. In: International conference on machine learning, pp 8599\u20138608. PMLR"},{"key":"10800_CR132","doi-asserted-by":"crossref","unstructured":"Preechakul K, Chatthee N, Wizadwongsa S, Suwajanakorn S (2022) Diffusion autoencoders: toward a meaningful and decodable representation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10619\u201310629","DOI":"10.1109\/CVPR52688.2022.01036"},{"key":"10800_CR133","unstructured":"Qian L, Wang M, Liu Y, Zhou H (2022) Diff-Glat: diffusion glancing transformer for parallel sequence to sequence learning. Preprint. arXiv:2212.10240"},{"key":"10800_CR134","unstructured":"Ramesh A, Dhariwal P, Nichol A, Chu C, Chen M (2022) Hierarchical text-conditional image generation with clip latents. Preprint. arXiv:2204.06125"},{"key":"10800_CR135","doi-asserted-by":"crossref","unstructured":"Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B (2022) High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10684\u201310695","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"10800_CR136","doi-asserted-by":"crossref","unstructured":"Ruan L, Ma Y, Yang H, He H, Liu B, Fu J, Yuan NJ, Jin Q, Guo B (2023) MM-Diffusion: learning multi-modal diffusion models for joint audio and video generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10219\u201310228","DOI":"10.1109\/CVPR52729.2023.00985"},{"key":"10800_CR137","doi-asserted-by":"crossref","unstructured":"Ruiz N, Li Y, Jampani V, Pritch Y, Rubinstein M, Aberman K (2023a) DreamBooth: fine tuning text-to-image diffusion models for subject-driven generation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 22500\u201322510","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"10800_CR138","doi-asserted-by":"crossref","unstructured":"Ruiz N, Li Y, Jampani V, Wei W, Hou T, Pritch Y, Wadhwa N, Rubinstein M, Aberman K (2023b) HyperDreamBooth: hypernetworks for fast personalization of text-to-image models. Preprint. arXiv:2307.06949","DOI":"10.1109\/CVPR52733.2024.00624"},{"key":"10800_CR139","unstructured":"Ryu D, Ye JC (2022) Pyramidal denoising diffusion probabilistic models. Preprint. arXiv:2208.01864"},{"key":"10800_CR140","unstructured":"Saharia C, Chan W, Saxena S, Li L, Whang J, Denton EL, Ghasemipour K, Gontijo Lopes R, Karagol Ayan B, Salimans T et al (2022a) Photorealistic text-to-image diffusion models with deep language understanding. In: Advances in neural information processing systems, vol 35, pp 36479\u201336494"},{"issue":"4","key":"10800_CR141","first-page":"4713","volume":"45","author":"C Saharia","year":"2022","unstructured":"Saharia C, Ho J, Chan W, Salimans T, Fleet DJ, Norouzi M (2022b) Image super-resolution via iterative refinement. IEEE Trans Pattern Anal Mach Intell 45(4):4713\u20134726","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"10800_CR143","unstructured":"Salimans T, Ho J (2021) Progressive distillation for fast sampling of diffusion models. In: International conference on learning representations"},{"key":"10800_CR142","unstructured":"Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Chen X (2016) Improved techniques for training GANs. In: Advances in neural information processing systems, vol 29"},{"key":"10800_CR144","unstructured":"Sanh V, Wolf T, Rush A (2020) Movement pruning: adaptive sparsity by fine-tuningg. In: Advances in neural information processing systems, vol 33, pp 20378\u201320389"},{"key":"10800_CR145","unstructured":"Sauer A, Lorenz D, Blattmann A, Rombach R (2023) Adversarial diffusion distillation. Preprint. arXiv:2311.17042"},{"key":"10800_CR146","unstructured":"Scarvelis C, Borde HSdO, Solomon J (2023) Closed-form diffusion models. Preprint. arXiv:2310.12395"},{"key":"10800_CR147","unstructured":"Sekhar\u00a0Sahoo S, Gokaslan A, De\u00a0Sa C, Kuleshov V (2023) Diffusion models with learned adaptive noise. Preprint. arXiv:2312.13236"},{"key":"10800_CR148","doi-asserted-by":"crossref","unstructured":"Shang S, Shan Z, Liu G, Zhang J (2023a) ResDiff: combining cnn and diffusion model for image super-resolution. Preprint. arXiv:2303.08714","DOI":"10.1609\/aaai.v38i8.28746"},{"key":"10800_CR149","doi-asserted-by":"crossref","unstructured":"Shang Y, Yuan Z, Xie B, Wu B, Yan Y (2023b) Post-training quantization on diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1972\u20131981","DOI":"10.1109\/CVPR52729.2023.00196"},{"key":"10800_CR150","unstructured":"Shao S, Dai X, Yin S, Li L, Chen H, Hu Y (2023) Catch-up distillation: you only need to train once for accelerating sampling. Preprint. arXiv:2305.10769"},{"key":"10800_CR151","doi-asserted-by":"crossref","unstructured":"Shen L, Yan J, Sun X, Li B, Pan Z (2023) Wavelet-based self-attention GAN with collaborative feature fusion for image inpainting. IEEE Trans Emerg Top Comput Intell 7:1651\u20131664","DOI":"10.1109\/TETCI.2023.3263200"},{"key":"10800_CR152","unstructured":"So J, Lee J, Ahn D, Kim H, Park E (2023) Temporal dynamic quantization for diffusion models. Preprint. arXiv:2306.02316"},{"key":"10800_CR154","unstructured":"Song Y, Dhariwal P (2023) Improved techniques for training consistency models. Preprint. arXiv:2310.14189"},{"key":"10800_CR156","unstructured":"Song J, Meng C, Ermon S (2020a) Denoising diffusion implicit models. In: International conference on learning representations"},{"key":"10800_CR157","unstructured":"Song Y, Sohl-Dickstein J, Kingma DP, Kumar A, Ermon S, Poole B (2020b) Score-based generative modeling through stochastic differential equations. In: International conference on learning representations"},{"issue":"2","key":"10800_CR153","doi-asserted-by":"publisher","first-page":"25","DOI":"10.3847\/1538-4365\/ac9a4d","volume":"263","author":"W Song","year":"2022","unstructured":"Song W, Ma W, Ma Y, Zhao X, Lin G (2022) Improving the spatial resolution of solar images based on an improved conditional denoising diffusion probability model. Astrophys J Suppl Ser 263(2):25","journal-title":"Astrophys J Suppl Ser"},{"key":"10800_CR155","unstructured":"Song Y, Dhariwal P, Chen M, Sutskever I (2023) Consistency models. Preprint. arXiv:2303.01469"},{"key":"10800_CR158","unstructured":"Starodubcev N, Fedorov A, Babenko A, Baranchuk D (2023) Your student is better than expected: adaptive teacher\u2013student collaboration for text-conditional diffusion models. Preprint. arXiv:2312.10835"},{"key":"10800_CR159","doi-asserted-by":"crossref","unstructured":"Strang G (1968) On the construction and comparison of difference schemes. SIAM J Numer Anal 5(3):506\u2013517","DOI":"10.1137\/0705041"},{"key":"10800_CR160","unstructured":"Sun W, Chen D, Wang C, Ye D, Feng Y, Chen C (2022) Accelerating diffusion sampling with classifier-based feature distillation. Preprint. arXiv:2211.12039"},{"key":"10800_CR161","doi-asserted-by":"crossref","unstructured":"Tang T, Chen Y, Du Y, Li J, Zhao WX, Wen J-R (2023) Learning to Imagine: visually-augmented natural language generation. Preprint. arXiv:2305.16944","DOI":"10.18653\/v1\/2023.acl-long.526"},{"key":"10800_CR162","unstructured":"Tsaban L, Passos A (2023) LEDITS: real image editing with DDPM inversion and semantic guidance. Preprint. arXiv:2307.00522"},{"issue":"1","key":"10800_CR163","first-page":"7184","volume":"17","author":"B Uria","year":"2016","unstructured":"Uria B, C\u00f4t\u00e9 M-A, Gregor K, Murray I, Larochelle H (2016) Neural autoregressive distribution estimation. J Mach Learn Res 17(1):7184\u20137220","journal-title":"J Mach Learn Res"},{"key":"10800_CR164","unstructured":"Vahdat A, Kreis K (2021) Kautz J Score-based generative modeling in latent space. In: Advances in neural information processing systems, vol 34, pp 11287\u201311302"},{"key":"10800_CR165","unstructured":"Voronov A, Khoroshikh M, Babenko A, Ryabinin M (2023) Is this loss informative? speeding up textual inversion with deterministic objective evaluation. Preprint. arXiv:2302.04841"},{"key":"10800_CR168","unstructured":"Wang X, Yan J-K, Cai J-Y, Deng J-H, Qin Q, Wang Q, Xiao H, Cheng Y, Ye P-F (2022a) Superresolution reconstruction of single image for latent features. Preprint. arXiv:2211.12845"},{"key":"10800_CR170","unstructured":"Wang T, Zhang T, Zhang B, Ouyang H, Chen D, Chen Q, Wen F (2022b) Pretraining is all you need for image-to-image translation. Preprint. arXiv:2205.12952"},{"key":"10800_CR171","unstructured":"Wang Z, Zheng H, He P, Chen W, Zhou M (2022c) Diffusion-Gan: training GANs with diffusion. In: The 11th International conference on learning representations"},{"key":"10800_CR166","doi-asserted-by":"crossref","unstructured":"Wang Z, Wang J, Liu Z, Qiu Q (2023a) Binary latent diffusion. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 22576\u201322585","DOI":"10.1109\/CVPR52729.2023.02162"},{"key":"10800_CR167","doi-asserted-by":"crossref","unstructured":"Wang C, Wang Z, Xu X, Tang Y, Zhou J, Lu J (2023b) Towards accurate data-free quantization for diffusion models. Preprint. arXiv:2305.18723","DOI":"10.1109\/CVPR52733.2024.01517"},{"key":"10800_CR169","doi-asserted-by":"crossref","unstructured":"Wang Y, Yang W, Chen X, Wang Y, Guo L, Chau L-P, Liu Z, Qiao Y, Kot AC, Wen B (2023c) SinSR: diffusion-based image super-resolution in a single step. Preprint. arXiv:2311.14760","DOI":"10.1109\/CVPR52733.2024.02437"},{"key":"10800_CR172","unstructured":"Watson D, Ho J, Norouzi M, Chan W (2021) Learning to efficiently sample from diffusion probabilistic models. Preprint. arXiv:2106.03802"},{"key":"10800_CR173","unstructured":"Wei X, Gong R, Li Y, Liu X, Yu F (2021) Qdrop: Randomly dropping quantization for extremely low-bit post-training quantization. In: International conference on learning representations"},{"key":"10800_CR174","unstructured":"Wizadwongsa S, Suwajanakorn S (2022) Accelerating guided diffusion sampling with splitting numerical methods. In: The Eleventh International Conference on Learning Representations"},{"key":"10800_CR175","doi-asserted-by":"crossref","unstructured":"Wortsman M, Ilharco G, Kim JW, Li M, Kornblith S, Roelofs R, Lopes R.G, Hajishirzi H, Farhadi A, Namkoong H et al (2022) Robust fine-tuning of zero-shot models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7959\u20137971","DOI":"10.1109\/CVPR52688.2022.00780"},{"key":"10800_CR176","doi-asserted-by":"crossref","unstructured":"Wu Q, Liu Y, Zhao H, Kale A, Bui T, Yu T, Lin Z, Zhang Y, Chang S (2023a) Uncovering the disentanglement capability in text-to-image diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1900\u20131910","DOI":"10.1109\/CVPR52729.2023.00189"},{"key":"10800_CR177","unstructured":"Wu Z, Zhou P, Kawaguchi K, Zhang H (2023b) Fast diffusion model. Preprint. arXiv:2306.06991"},{"key":"10800_CR178","unstructured":"Xia W, Cong W, Wang G (2022) Patch-based denoising diffusion probabilistic model for sparse-view CT reconstruction. Preprint. arXiv:2211.10388"},{"key":"10800_CR182","doi-asserted-by":"crossref","unstructured":"Xia M, Shen Y, Lei C, Zhou Y, Yi R, Zhao D, Wang W, Liu Y-j (2023a) Towards more accurate diffusion model acceleration with a timestep aligner. Preprint. arXiv:2310.09469","DOI":"10.1109\/CVPR52733.2024.00548"},{"key":"10800_CR183","doi-asserted-by":"crossref","unstructured":"Xia B, Zhang Y, Wang S, Wang Y, Wu X, Tian Y, Yang W, Van\u00a0Gool L (2023b) DiffIR: efficient diffusion model for image restoration. Preprint. arXiv:2303.09472","DOI":"10.1109\/ICCV51070.2023.01204"},{"key":"10800_CR179","unstructured":"Xiao Z, Kreis K, Vahdat A (2021) Tackling the generative learning trilemma with denoising diffusion gans. In: International conference on learning representations"},{"key":"10800_CR180","unstructured":"Xiao J, Yin M, Gong Y, Zang X, Ren J, Yuan B (2023a) COMCAT: towards efficient compression and customization of attention-based vision models. Preprint. arXiv:2305.17235"},{"key":"10800_CR181","doi-asserted-by":"crossref","unstructured":"Xiao Y, Yuan Q, Jiang K, He J, Jin X, Zhang L (2023b) EDiffSR: an efficient diffusion probabilistic model for remote sensing image super-resolution. IEEE Trans Geosci Remot Sens 62:5601514","DOI":"10.1109\/TGRS.2023.3341437"},{"key":"10800_CR184","doi-asserted-by":"crossref","unstructured":"Xie E, Yao L, Shi H, Liu Z, Zhou D, Liu Z, Li J, Li Z (2023) DiffFit: unlocking transferability of large diffusion models via simple parameter-efficient fine-tuning. Preprint. arXiv:2304.06648","DOI":"10.1109\/ICCV51070.2023.00390"},{"key":"10800_CR185","unstructured":"Xu Y, Gong M, Xie S, Wei W, Grundmann M, Hou T et al (2023) Semi-implicit denoising diffusion models (SIDDMs). Preprint. arXiv:2306.12511"},{"key":"10800_CR186","unstructured":"Yang S, Chen Y, Wang L, Liu S, Chen Y (2023a) Denoising diffusion step-aware models. Preprint. arXiv:2310.03337"},{"key":"10800_CR187","doi-asserted-by":"crossref","unstructured":"Yang B, Gu S, Zhang B, Zhang T, Chen X, Sun X, Chen D, Wen F (2023b) Paint by example: exemplar-based image editing with diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18381\u201318391","DOI":"10.1109\/CVPR52729.2023.01763"},{"key":"10800_CR188","doi-asserted-by":"crossref","unstructured":"Yang X, Zhou D, Feng J, Wang X (2023c) Diffusion probabilistic model made slim. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 22552\u201322562","DOI":"10.1109\/CVPR52729.2023.02160"},{"key":"10800_CR189","doi-asserted-by":"crossref","unstructured":"Yin G, Wang W, Yuan Z, Han C, Ji W, Sun S, Wang C (2022) Content-variant reference image quality assessment via knowledge distillation. In: Proceedings of the AAAI conference on artificial intelligence, vol 36, pp 3134\u20133142","DOI":"10.1609\/aaai.v36i3.20221"},{"key":"10800_CR190","unstructured":"Yin T, Gharbi M, Zhang R, Shechtman E, Durand F, Freeman WT, Park T (2023) One-step diffusion with distribution matching distillation. Preprint. arXiv:2311.18828"},{"key":"10800_CR191","doi-asserted-by":"crossref","unstructured":"Youn J, Song J, Kim H-S, Bahk S (2022) Bitwidth-adaptive quantization-aware neural network training: a meta-learning approach. In: European conference on computer vision. Springer, Cham, pp 208\u2013224","DOI":"10.1007\/978-3-031-19775-8_13"},{"key":"10800_CR193","unstructured":"Yu F, Seff A, Zhang Y, Song S, Funkhouser T, Xiao J (2015) LSUN: construction of a large-scale image dataset using deep learning with humans in the loop. Preprint. arXiv:1506.03365"},{"key":"10800_CR192","unstructured":"Yue Z, Wang J, Loy CC (2023) ResShift: efficient diffusion model for image super-resolution by residual shifting. Preprint. arXiv:2307.12348"},{"key":"10800_CR194","unstructured":"Yu H, Shen L, Huang J, Zhou M, Li H, Zhao F (2023a) Debias the training of diffusion models. Preprint. arXiv:2310.08442"},{"key":"10800_CR195","doi-asserted-by":"crossref","unstructured":"Yu S, Sohn K, Kim S, Shin J (2023b) Video probabilistic diffusion models in projected latent space. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18456\u201318466","DOI":"10.1109\/CVPR52729.2023.01770"},{"key":"10800_CR196","unstructured":"Zhang Q (2021) Diffusion normalizing flow. In: Advances in neural information processing systems, vol 34, pp 16280\u201316291"},{"key":"10800_CR199","unstructured":"Zhang Q, Chen Y (2022) Fast sampling of diffusion models with exponential integrator. In: The 11th International conference on learning representations"},{"key":"10800_CR198","doi-asserted-by":"crossref","unstructured":"Zhang L, Agrawala M (2023) Adding conditional control to text-to-image diffusion models. Preprint. arXiv:2302.05543","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"10800_CR203","doi-asserted-by":"crossref","unstructured":"Zhang K, Liang J, Van\u00a0Gool L, Timofte R (2021) Designing a practical degradation model for deep blind image super-resolution. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 4791\u20134800","DOI":"10.1109\/ICCV48922.2021.00475"},{"key":"10800_CR197","unstructured":"Zhang Z, Zhao Z, Lin Z (2022a) Unsupervised representation learning from pre-trained diffusion probabilistic models. In: Advances in neural information processing systems, vol 35, pp 22117\u201322130"},{"key":"10800_CR200","doi-asserted-by":"crossref","unstructured":"Zhang L, Chen X, Tu X, Wan P, Xu N, Ma K (2022b) Wavelet knowledge distillation: towards efficient image-to-image translation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12464\u201312474","DOI":"10.1109\/CVPR52688.2022.01214"},{"key":"10800_CR201","unstructured":"Zhang Z, Ehinger KA, Drummond T (2023a) Improving denoising diffusion models via simultaneous estimation of image and noise. Preprint. arXiv:2310.17167"},{"key":"10800_CR202","doi-asserted-by":"crossref","unstructured":"Zhang H, Feng R, Yang Z, Huang L, Liu Y, Zhang Y, Shen Y, Zhao D, Zhou J, Cheng F (2023b) Dimensionality-varying diffusion process. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14307\u201314316","DOI":"10.1109\/CVPR52729.2023.01375"},{"key":"10800_CR204","unstructured":"Zhang L, Ma H, Zhu X, Feng J (2023c) Preconditioned score-based generative models. Preprint. arXiv:2302.06504"},{"key":"10800_CR205","unstructured":"Zhang H, Wang Z, Wu Z, Jiang Y-G (2023d) DiffusionAD: denoising diffusion for anomaly detection. Preprint. arXiv:2303.08730"},{"key":"10800_CR206","unstructured":"Zhang K, Yang X, Wang WY, Li L (2023e) ReDi: efficient learning-free diffusion inference via trajectory retrieval. Preprint. arXiv:2302.02285"},{"key":"10800_CR207","unstructured":"Zhao W, Bai L, Rao Y, Zhou J, Lu J (2023a) UniPC: a unified predictor-corrector framework for fast sampling of diffusion models. Preprint. arXiv:2302.04867"},{"key":"10800_CR208","doi-asserted-by":"crossref","unstructured":"Zhao K, Hung ALY, Pang K, Zheng H, Sung K (2023b) PartDiff: image super-resolution with partial diffusion models. Preprint. arXiv:2307.11926","DOI":"10.1109\/TMI.2024.3483109"},{"key":"10800_CR209","doi-asserted-by":"publisher","unstructured":"Zhao C, Yang P, Zhou F, Yue G, Wang S, Wu H, Chen G, Wang T, Lei B (2023c) MHW-GAN: multidiscriminator hierarchical wavelet generative adversarial network for multimodal image fusion. IEEE Trans Neural Netw Learn Syst. https:\/\/doi.org\/10.1109\/TNNLS.2023.3271059","DOI":"10.1109\/TNNLS.2023.3271059"},{"key":"10800_CR210","unstructured":"Zheng H, Nie W, Vahdat A, Anandkumar A (2023a) Fast training of diffusion models with masked transformers. Preprint. arXiv:2306.09305"},{"key":"10800_CR211","unstructured":"Zheng H, Nie W, Vahdat A, Azizzadenesheli K, Anandkumar A (2023b) Fast sampling of diffusion models via operator learning. In: International conference on machine learning, pp 42390\u201342402. PMLR"},{"key":"10800_CR212","unstructured":"Zheng H, Wang Z, Yuan J, Ning G, He P, You Q, Yang H, Zhou M (2023c) Learning stackable and skippable LEGO bricks for efficient, reconfigurable, and variable-resolution diffusion modeling. Preprint. arXiv:2310.06389"},{"key":"10800_CR213","doi-asserted-by":"crossref","unstructured":"Zhou Z, Chen D, Wang C, Chen C (2023a) Fast ODE-based sampling for diffusion models in around 5 steps. Preprint. arXiv:2312.00094","DOI":"10.1109\/CVPR52733.2024.00743"},{"key":"10800_CR214","doi-asserted-by":"crossref","unstructured":"Zhou D, Yang Z, Yang Y (2023b) Pyramid diffusion models for low-light image enhancement. Preprint. arXiv:2305.10028","DOI":"10.24963\/ijcai.2023\/199"},{"key":"10800_CR215","unstructured":"Zhu J, Ma H, Chen J, Yuan J (2023) DomainStudio: fine-tuning diffusion models for domain-driven image generation using limited data. Preprint. arXiv:2306.14153"}],"container-title":["Artificial Intelligence Review"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10462-024-10800-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10462-024-10800-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10462-024-10800-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T20:22:09Z","timestamp":1732134129000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10462-024-10800-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,31]]},"references-count":215,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2024,6]]}},"alternative-id":["10800"],"URL":"https:\/\/doi.org\/10.1007\/s10462-024-10800-8","relation":{},"ISSN":["1573-7462"],"issn-type":[{"value":"1573-7462","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,31]]},"assertion":[{"value":"16 May 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 May 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"161"}}