{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T05:05:18Z","timestamp":1765343118313,"version":"3.46.0"},"publisher-location":"New York, NY, USA","reference-count":51,"publisher":"ACM","funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176108"],"award-info":[{"award-number":["62176108"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,10,27]]},"DOI":"10.1145\/3746027.3755608","type":"proceedings-article","created":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T07:27:39Z","timestamp":1761377259000},"page":"10370-10378","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Frequency Regulation for Exposure Bias Mitigation in Diffusion Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-5121-5816","authenticated-orcid":false,"given":"Meng","family":"Yu","sequence":"first","affiliation":[{"name":"School of Information Science and Engineering, Lanzhou University, Lanzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8000-5682","authenticated-orcid":false,"given":"Kun","family":"Zhan","sequence":"additional","affiliation":[{"name":"School of Information Science and Engineering, Lanzhou University, Lanzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Fan Bao Chongxuan Li Jiacheng Sun Jun Zhu and Bo Zhang. 2022a. Estimating the Optimal Covariance with Imperfect Mean in Diffusion Probabilistic Models. In ICML."},{"key":"e_1_3_2_1_2_1","volume-title":"International Conference on Learning Representations.","author":"Bao Fan","year":"2022","unstructured":"Fan Bao, Chongxuan Li, Jun Zhu, and Bo Zhang. 2022b. Analytic-DPM: an Analytic Estimate of the Optimal Reverse Variance in Diffusion Probabilistic Models. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_3_1","volume-title":"A downsampled variant of ImageNet as an alternative to the CIFAR datasets. arXiv preprint arXiv:1707.08819","author":"Chrabaszcz Patryk","year":"2017","unstructured":"Patryk Chrabaszcz, Ilya Loshchilov, and Frank Hutter. 2017. A downsampled variant of ImageNet as an alternative to the CIFAR datasets. arXiv preprint arXiv:1707.08819 (2017)."},{"key":"e_1_3_2_1_4_1","unstructured":"Prafulla Dhariwal and Alexander Nichol. 2021. Diffusion models beat GANs on image synthesis. In NeurIPS."},{"key":"e_1_3_2_1_5_1","volume-title":"Genie: Higher-order denoising diffusion solvers. In NeurIPS.","author":"Dockhorn Tim","year":"2022","unstructured":"Tim Dockhorn, Arash Vahdat, and Karsten Kreis. 2022. Genie: Higher-order denoising diffusion solvers. In NeurIPS."},{"key":"e_1_3_2_1_6_1","volume-title":"Amit Bermano, and Daniel Cohen-Or.","author":"Gal Rinon","year":"2021","unstructured":"Rinon Gal, Dana Cohen Hochberg, Amit Bermano, and Daniel Cohen-Or. 2021. Swagan: A style-based wavelet-driven generative model. ACM Transactions on Graphics (2021)."},{"key":"e_1_3_2_1_7_1","unstructured":"Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial nets. In NeurIPS."},{"key":"e_1_3_2_1_8_1","volume-title":"An introduction to wavelets","author":"Graps Amara","year":"1995","unstructured":"Amara Graps. 1995. An introduction to wavelets. IEEE Computational Science and Engineering (1995)."},{"key":"e_1_3_2_1_9_1","volume-title":"Valentin De Bortoli, and Stephane Mallat","author":"Guth Florentin","year":"2022","unstructured":"Florentin Guth, Simon Coste, Valentin De Bortoli, and Stephane Mallat. 2022. Wavelet score-based generative modeling. In NeurIPS."},{"key":"e_1_3_2_1_10_1","unstructured":"Martin Heusel Hubert Ramsauer Thomas Unterthiner Bernhard Nessler and Sepp Hochreiter. 2017. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In NeurIPS."},{"key":"e_1_3_2_1_11_1","unstructured":"Jonathan Ho Ajay Jain and Pieter Abbeel. 2020. Denoising Diffusion Probabilistic Models. In NeurIPS."},{"key":"e_1_3_2_1_12_1","unstructured":"Tero Karras Miika Aittala Timo Aila and Samuli Laine. 2022. Elucidating the Design Space of Diffusion-Based Generative Models. In NeurIPS."},{"key":"e_1_3_2_1_13_1","volume-title":"Wanmo Kang, and Il-Chul Moon.","author":"Kim Dongjun","year":"2023","unstructured":"Dongjun Kim, Yeongmin Kim, Se Jung Kwon, Wanmo Kang, and Il-Chul Moon. 2023. Refining Generative Process with Discriminator Guidance in Score-based Diffusion Models. In ICML."},{"key":"e_1_3_2_1_14_1","unstructured":"Alex Krizhevsky Geoffrey Hinton et al. 2009. Learning multiple layers of features from tiny images. (2009)."},{"key":"e_1_3_2_1_15_1","volume-title":"Wavelet transform-assisted adaptive generative modeling for colorization","author":"Li Jin","year":"2022","unstructured":"Jin Li, Wanyun Li, Zichen Xu, Yuhao Wang, and Qiegen Liu. 2022. Wavelet transform-assisted adaptive generative modeling for colorization. IEEE Transactions on Multimedia (2022)."},{"key":"e_1_3_2_1_16_1","unstructured":"Mingxiao Li Tingyu Qu Ruicong Yao Wei Sun and Marie-Francine Moens. 2024. Alleviating Exposure Bias in Diffusion Models through Sampling with Shifted Time Steps. In ICLR."},{"key":"e_1_3_2_1_17_1","unstructured":"Yangming Li and Mihaela van der Schaar. 2024. On Error Propagation of Diffusion Models. In ICLR."},{"key":"e_1_3_2_1_18_1","volume-title":"Instaflow: One step is enough for high-quality diffusion-based text-to-image generation. In ICLR.","author":"Liu Xingchao","year":"2023","unstructured":"Xingchao Liu, Xiwen Zhang, Jianzhu Ma, Jian Peng, et al., 2023. Instaflow: One step is enough for high-quality diffusion-based text-to-image generation. In ICLR."},{"key":"e_1_3_2_1_19_1","unstructured":"Cheng Lu and Yang Song. 2025. Simplifying Stabilizing and Scaling Continuous-time Consistency Models. In ICLR."},{"key":"e_1_3_2_1_20_1","unstructured":"Cheng Lu Yuhao Zhou Fan Bao Jianfei Chen Chongxuan Li and Jun Zhu. 2022. DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps. In NeurIPS."},{"key":"e_1_3_2_1_21_1","volume-title":"Knowledge distillation in iterative generative models for improved sampling speed. arXiv preprint arXiv:2101.02388","author":"Luhman Eric","year":"2021","unstructured":"Eric Luhman and Troy Luhman. 2021. Knowledge distillation in iterative generative models for improved sampling speed. arXiv preprint arXiv:2101.02388 (2021)."},{"key":"e_1_3_2_1_22_1","volume-title":"A theory for multiresolution signal decomposition: the wavelet representation","author":"Mallat Stephane G","year":"1989","unstructured":"Stephane G Mallat. 1989. A theory for multiresolution signal decomposition: the wavelet representation. IEEE Transactions on Pattern Analysis and Machine Intelligence (1989)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Chenlin Meng Robin Rombach Ruiqi Gao Diederik Kingma Stefano Ermon Jonathan Ho and Tim Salimans. 2023. On distillation of guided diffusion models. In CVPR.","DOI":"10.1109\/CVPR52729.2023.01374"},{"key":"e_1_3_2_1_24_1","unstructured":"Alexander Quinn Nichol and Prafulla Dhariwal. 2021. Improved denoising diffusion probabilistic models. In ICLR."},{"key":"e_1_3_2_1_25_1","volume-title":"Albert Ali Salah, and Itir Onal Ertugrul","author":"Ning Mang","year":"2024","unstructured":"Mang Ning, Mingxiao Li, Jianlin Su, Albert Ali Salah, and Itir Onal Ertugrul. 2024. Elucidating the Exposure Bias in Diffusion Models. In ICLR."},{"key":"e_1_3_2_1_26_1","unstructured":"Mang Ning Enver Sangineto Angelo Porrello Simone Calderara and Rita Cucchiara. 2023. Input Perturbation Reduces Exposure Bias in Diffusion Models. In ICML."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Hao Phung Quan Dao and Anh Tran. 2023. Wavelet diffusion models are fast and scalable image generators. In CVPR.","DOI":"10.1109\/CVPR52729.2023.00983"},{"key":"e_1_3_2_1_28_1","first-page":"4667","article-title":"Multi-Step Denoising Scheduled Sampling","author":"Ren Zhiyao","year":"2024","unstructured":"Zhiyao Ren, Yibing Zhan, Liang Ding, Gaoang Wang, Chaoyue Wang, Zhongyi Fan, and Dacheng Tao. 2024. Multi-Step Denoising Scheduled Sampling: Towards Alleviating Exposure Bias for Diffusion Models. In AAAI. 4667-4675.","journal-title":"Towards Alleviating Exposure Bias for Diffusion Models. In AAAI."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2022. High-resolution image synthesis with latent diffusion models. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_2_1_30_1","unstructured":"Tim Salimans Ian Goodfellow Wojciech Zaremba Vicki Cheung Alec Radford and Xi Chen. 2016. Improved techniques for training GANs. In NeurIPS."},{"key":"e_1_3_2_1_31_1","unstructured":"Tim Salimans and Jonathan Ho. 2022. Progressive Distillation for Fast Sampling of Diffusion Models. In ICLR."},{"key":"e_1_3_2_1_32_1","first-page":"2256","article-title":"Deep unsupervised learning using nonequilibrium thermodynamics","author":"Sohl-Dickstein Jascha","year":"2015","unstructured":"Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. 2015. Deep unsupervised learning using nonequilibrium thermodynamics. In ICML. 2256-2265.","journal-title":"ICML."},{"key":"e_1_3_2_1_33_1","unstructured":"Jiaming Song Chenlin Meng and Stefano Ermon. 2021a. Denoising Diffusion Implicit Models. In ICLR."},{"key":"e_1_3_2_1_34_1","unstructured":"Yang Song and Prafulla Dhariwal. 2024. Improved Techniques for Training Consistency Models. In ICLR."},{"key":"e_1_3_2_1_35_1","unstructured":"Yang Song Prafulla Dhariwal Mark Chen and Ilya Sutskever. 2023. Consistency models. In ICML."},{"key":"e_1_3_2_1_36_1","unstructured":"Yang Song Jascha Sohl-Dickstein Diederik P Kingma Abhishek Kumar Stefano Ermon and Ben Poole. 2021b. Score-Based Generative Modeling through Stochastic Differential Equations. In ICLR."},{"key":"e_1_3_2_1_37_1","unstructured":"Arash Vahdat Karsten Kreis and Jan Kautz. 2021. Score-based generative modeling in latent space. NeurIPS."},{"key":"e_1_3_2_1_38_1","unstructured":"Zhe Wang Ziqiu Chi Yanbing Zhang et al. 2022. FreGAN: Exploiting frequency components for training GANs under limited data. NeurIPS (2022)."},{"key":"e_1_3_2_1_39_1","unstructured":"Yilun Xu Ziming Liu Max Tegmark and Tommi Jaakkola. 2022. Poisson flow generative models. In NeurIPS."},{"key":"e_1_3_2_1_40_1","volume-title":"PFGM: Unlocking the potential of physics-inspired generative models. In ICML.","author":"Xu Yilun","year":"2023","unstructured":"Yilun Xu, Ziming Liu, Yonglong Tian, Shangyuan Tong, Max Tegmark, and Tommi Jaakkola. 2023. PFGM: Unlocking the potential of physics-inspired generative models. In ICML."},{"key":"e_1_3_2_1_41_1","volume-title":"Wavegan: Frequency-aware gan for high-fidelity few-shot image generation. In ECCV.","author":"Yang Mengping","year":"2022","unstructured":"Mengping Yang, Zhe Wang, Ziqiu Chi, and Wenyi Feng. 2022. Wavegan: Frequency-aware gan for high-fidelity few-shot image generation. In ECCV."},{"key":"e_1_3_2_1_42_1","unstructured":"Yuzhe YAO Jun Chen Zeyi Huang Haonan Lin Mengmeng Wang Guang Dai and Jingdong Wang. 2025. Manifold Constraint Reduces Exposure Bias in Accelerated Diffusion Sampling. In ICLR."},{"key":"e_1_3_2_1_43_1","volume-title":"Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365","author":"Yu Fisher","year":"2015","unstructured":"Fisher Yu, Ari Seff, Yinda Zhang, Shuran Song, Thomas Funkhouser, and Jianxiong Xiao. 2015. Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365 (2015)."},{"key":"e_1_3_2_1_44_1","unstructured":"Meng Yu and Kun Zhan. 2025. Bias Mitigation in Graph Diffusion Models. In ICLR."},{"key":"e_1_3_2_1_45_1","volume-title":"Styleswin: Transformer-based gan for high-resolution image generation. In CVPR.","author":"Zhang Bowen","year":"2022","unstructured":"Bowen Zhang, Shuyang Gu, Bo Zhang, Jianmin Bao, Dong Chen, Fang Wen, Yong Wang, and Baining Guo. 2022. Styleswin: Transformer-based gan for high-resolution image generation. In CVPR."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"crossref","unstructured":"Guoqiang Zhang Kenta Niwa and W Bastiaan Kleijn. 2023. Lookahead diffusion probabilistic models for refining mean estimation. In CVPR.","DOI":"10.1109\/CVPR52729.2023.00143"},{"key":"e_1_3_2_1_47_1","unstructured":"Junyu Zhang Daochang Liu Eunbyung Park Shichao Zhang and Chang Xu. 2025. Anti-Exposure Bias in Diffusion Models. In ICLR."},{"key":"e_1_3_2_1_48_1","unstructured":"Qinsheng Zhang and Yongxin Chen. 2023. Fast Sampling of Diffusion Models with Exponential Integrator. In ICLR."},{"key":"e_1_3_2_1_49_1","volume-title":"Unipc: A unified predictor-corrector framework for fast sampling of diffusion models. In NeurIPS.","author":"Zhao Wenliang","year":"2024","unstructured":"Wenliang Zhao, Lujia Bai, Yongming Rao, Jie Zhou, and Jiwen Lu. 2024. Unipc: A unified predictor-corrector framework for fast sampling of diffusion models. In NeurIPS."},{"key":"e_1_3_2_1_50_1","unstructured":"Hongkai Zheng Weili Nie Arash Vahdat Kamyar Azizzadenesheli and Anima Anandkumar. 2023. Fast sampling of diffusion models via operator learning. In ICML."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"crossref","unstructured":"Zhenyu Zhou Defang Chen Can Wang and Chun Chen. 2024. Fast ode-based sampling for diffusion models in around 5 steps. In CVPR.","DOI":"10.1109\/CVPR52733.2024.00743"}],"event":{"name":"MM '25: The 33rd ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Dublin Ireland","acronym":"MM '25"},"container-title":["Proceedings of the 33rd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746027.3755608","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T05:01:24Z","timestamp":1765342884000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746027.3755608"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"references-count":51,"alternative-id":["10.1145\/3746027.3755608","10.1145\/3746027"],"URL":"https:\/\/doi.org\/10.1145\/3746027.3755608","relation":{},"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"2025-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}