{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T11:04:48Z","timestamp":1773486288601,"version":"3.50.1"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,2,17]],"date-time":"2026-02-17T00:00:00Z","timestamp":1771286400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,17]],"date-time":"2026-02-17T00:00:00Z","timestamp":1771286400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"crossref","award":["6256205"],"award-info":[{"award-number":["6256205"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"National Science Foundation of Ningxia","award":["2025AAC020023"],"award-info":[{"award-number":["2025AAC020023"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Pattern Anal Applic"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s10044-026-01632-z","type":"journal-article","created":{"date-parts":[[2026,2,17]],"date-time":"2026-02-17T00:25:19Z","timestamp":1771287919000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Fast and faithful: accelerating data-free knowledge distillation via confidence-aware adaptive diffusion"],"prefix":"10.1007","volume":"29","author":[{"given":"Chenyang","family":"Jiang","sequence":"first","affiliation":[]},{"given":"Zhendong","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shuai","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,17]]},"reference":[{"key":"1632_CR1","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1632_CR2","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, et al. (2020) An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929"},{"key":"1632_CR3","unstructured":"Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531"},{"key":"1632_CR4","unstructured":"Lopes RG, Fenu S, Starner T (2017) Data-free knowledge distillation for deep neural networks. In: NIPS workshop on machine learning for health"},{"key":"1632_CR5","doi-asserted-by":"crossref","unstructured":"Chen H, Wang Y, Xu C, Yang Z, Liu C, Shi B, Xu C, Xu C, Tian Q (2019) Data-free learning of student networks. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 3514\u20133522","DOI":"10.1109\/ICCV.2019.00361"},{"key":"1632_CR6","unstructured":"Fang G, Song J, Wang X, Song M (2019) Data-free adversarial distillation. arXiv preprint arXiv:1912.11006"},{"key":"1632_CR7","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. Adv Neural Inform Process Syst 27"},{"key":"1632_CR8","doi-asserted-by":"crossref","unstructured":"Yin H, Molchanov P, Alvarez JM, Li Z, Mallya A, Hoiem D, Jha NK, Kautz J (2020) Dreaming to distill: data-free knowledge transfer via deepinversion. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 8715\u20138724","DOI":"10.1109\/CVPR42600.2020.00874"},{"key":"1632_CR9","unstructured":"Choi K, et al. (2020) Data-free knowledge distillation with soft targeted transfer. In: Proceedings of the IEEE\/CVF International conference on computer vision workshops"},{"key":"1632_CR10","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. Adv Neural Inf Process Syst 33:6840\u20136851","journal-title":"Adv Neural Inf Process Syst"},{"key":"1632_CR11","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal P, Nichol A (2021) Diffusion models beat gans on image synthesis. Adv Neural Inf Process Syst 34:8780\u20138794","journal-title":"Adv Neural Inf Process Syst"},{"key":"1632_CR12","doi-asserted-by":"crossref","unstructured":"Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B (2022) High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 10684\u201310695","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"1632_CR13","unstructured":"Song J, Meng C, Ermon S (2020) Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502"},{"key":"1632_CR14","unstructured":"Nichol A, Dhariwal P (2021) Improved denoising diffusion probabilistic models. In: International conference on machine learning, pp. 8162\u20138171. PMLR"},{"key":"1632_CR15","first-page":"659","volume":"36","author":"G Fang","year":"2022","unstructured":"Fang G, Ma K, Song J, Mi MB, Wang X (2022) Up to 100x faster data-free knowledge distillation. Proc AAAI Conf Artif Intell 36:659\u2013667","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"1632_CR16","unstructured":"Nayak GK, Mopuri KR, Shajil V, Radhakrishnan VB, Chakraborty A (2019) Zero-shot knowledge distillation in deep networks. In: International conference on machine learning, pp. 4743\u20134751"},{"key":"1632_CR17","doi-asserted-by":"crossref","unstructured":"Fang G, Song J, Wang X, Song M (2021) Contrastive model inversion for data-free knowledge distillation. In: Proceedings of the international joint conference on artificial intelligence, pp. 2374\u20132380","DOI":"10.24963\/ijcai.2021\/327"},{"key":"1632_CR18","doi-asserted-by":"crossref","unstructured":"Tran M-T, Le T, Le X-M, Harandi M, Tran QH, Phung D (2024) Nayer: Noisy layer data generation for efficient and effective data-free knowledge distillation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 23859\u201323869","DOI":"10.1109\/CVPR52733.2024.02252"},{"key":"1632_CR19","unstructured":"Ye J, Ji Y, Wang X, Song J, Song M (2023) Small scale data-free knowledge distillation. In: Proceedings of the international joint conference on artificial intelligence (IJCAI)"},{"key":"1632_CR20","unstructured":"He X, Wang J, Cheng Z, Su Y, Huang S, Lu G (2025) Prism: precision-recall informed data-free knowledge distillation via generative diffusion. arXiv preprint arXiv:2509.16897"},{"key":"1632_CR21","doi-asserted-by":"crossref","unstructured":"Qi X, Li R, Peng L, Ling Q, Yu J, Chen Z, Chang P, Han M, Xiao J (2025) Data-free knowledge distillation with diffusion models. arXiv preprint arXiv:2504.00870. ICME 2025","DOI":"10.1109\/ICME59968.2025.11209042"},{"key":"1632_CR22","doi-asserted-by":"crossref","unstructured":"Xiang Q, Zhang M, Shang Y, Wu J, Yan Y, Nie L (2024) Dkdm: data-free knowledge distillation for diffusion models with any architecture. arXiv preprint arXiv:2409.03550. CVPR 2025","DOI":"10.1109\/CVPR52734.2025.00281"},{"key":"1632_CR23","unstructured":"Sohl-Dickstein J, Weiss E, Maheswaranathan N, Ganguli S (2015) Deep unsupervised learning using nonequilibrium thermodynamics. In: International conference on machine learning, pp. 2256\u20132265"},{"key":"1632_CR24","unstructured":"Song Y, Sohl-Dickstein J, Kingma DP, Kumar A, Ermon S, Poole B (2021) Score-based generative modeling through stochastic differential equations. In: International conference on learning representations"},{"key":"1632_CR25","unstructured":"Ho J, Salimans T (2022) Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598"},{"key":"1632_CR26","unstructured":"Liu L, Ren Y, Lin Z, Zhao Z (2022) Pseudo numerical methods for diffusion models on manifolds. In: International conference on learning representations"},{"key":"1632_CR27","first-page":"5775","volume":"35","author":"C Lu","year":"2022","unstructured":"Lu C, Zhou Y, Bao F, Chen J, Li C, Zhu J (2022) Dpm-solver: a fast ode solver for diffusion probabilistic model sampling in around 10 steps. Adv Neural Inform Process Syst 35: 5775\u20135787","journal-title":"Adv Neural Inform Process Syst"},{"key":"1632_CR28","unstructured":"Luhman E, Luhman T (2021) Knowledge distillation in iterative generative models for improved sampling speed. arXiv preprint arXiv:2101.02388"},{"key":"1632_CR29","unstructured":"Krizhevsky A, Hinton G, et al. (2009) Learning multiple layers of features from tiny images. Master\u2019s thesis, Department of Computer Science, University of Toronto"},{"issue":"7","key":"1632_CR30","first-page":"3","volume":"7","author":"Y Le","year":"2015","unstructured":"Le Y, Yang X (2015) Tiny imagenet visual recognition challenge. CS 231N 7(7):3","journal-title":"CS 231N"},{"key":"1632_CR31","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1632_CR32","unstructured":"Howard J (2019) ImageNette: A smaller subset of 10 easily classified classes from Imagenet. https:\/\/github.com\/fastai\/imagenette"},{"key":"1632_CR33","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: International conference on learning representations"},{"key":"1632_CR34","doi-asserted-by":"crossref","unstructured":"Zagoruyko S, Komodakis N (2016) Wide residual networks. In: Proceedings of the British machine vision conference (BMVC)","DOI":"10.5244\/C.30.87"},{"key":"1632_CR35","doi-asserted-by":"crossref","unstructured":"Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4510\u20134520","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1632_CR36","doi-asserted-by":"crossref","unstructured":"Li Y, Kim Y, Lee D, Kundu S, Panda P (2024) Genq: Quantization in low data regimes with generative synthetic data. In: European conference on computer vision, pp. 216\u2013235. Springer","DOI":"10.1007\/978-3-031-72624-8_13"},{"key":"1632_CR37","unstructured":"Li Z, Li Y, Zhao P, Song R, Li X, Yang J (2023) Is synthetic data from diffusion models ready for knowledge distillation? arXiv preprint arXiv:2305.12954"},{"key":"1632_CR38","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2024.3442295","author":"J Xue","year":"2024","unstructured":"Xue J, Zhao Y, Wu T, Chan JC-W (2024) Tensor convolution-like low-rank dictionary for high-dimensional image representation. IEEE Trans Circuits Syst Video Technol. https:\/\/doi.org\/10.1109\/TCSVT.2024.3442295","journal-title":"IEEE Trans Circuits Syst Video Technol"}],"container-title":["Pattern Analysis and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-026-01632-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10044-026-01632-z","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-026-01632-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T10:38:31Z","timestamp":1773484711000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10044-026-01632-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,17]]},"references-count":38,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["1632"],"URL":"https:\/\/doi.org\/10.1007\/s10044-026-01632-z","relation":{},"ISSN":["1433-7541","1433-755X"],"issn-type":[{"value":"1433-7541","type":"print"},{"value":"1433-755X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,17]]},"assertion":[{"value":"18 December 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"52"}}