{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T18:56:15Z","timestamp":1757616975864,"version":"3.44.0"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"23","license":[{"start":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T00:00:00Z","timestamp":1726272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,14]],"date-time":"2024-09-14T00:00:00Z","timestamp":1726272000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61972096","61771140","61872088","61872090"],"award-info":[{"award-number":["61972096","61771140","61872088","61872090"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005270","name":"Fujian Provincial Department of Science and Technology","doi-asserted-by":"publisher","award":["2022H60250"],"award-info":[{"award-number":["2022H60250"]}],"id":[{"id":"10.13039\/501100005270","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-20193-z","type":"journal-article","created":{"date-parts":[[2024,9,13]],"date-time":"2024-09-13T22:01:55Z","timestamp":1726264915000},"page":"26935-26962","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DE-DFKD: diversity enhancing data-free knowledge distillation"],"prefix":"10.1007","volume":"84","author":[{"given":"Yanni","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2606-5406","authenticated-orcid":false,"given":"Ayong","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Qiulin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yuexin","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jianwei","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,14]]},"reference":[{"issue":"7","key":"20193_CR1","first-page":"38","volume":"14","author":"G Hinton","year":"2015","unstructured":"Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38\u201339","journal-title":"Comput Sci"},{"key":"20193_CR2","unstructured":"Romero A, Ballas N, Kahou SE, Chassang A, Gatta C, Bengio Y (2015) Fitnets: hints for thin deep nets, 2015. Paper presented at International Conference on Learning Representations (ICLR), pp 7\u20139."},{"key":"20193_CR3","unstructured":"Lopes RG, Fenu S, Starner T (2017) Data-free knowledge distillation for deep neural networks, 2017. Paper presented at Neural Information Processing Systems (NeurIPS), pp 1\u20138"},{"issue":"2","key":"20193_CR4","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/s10462-023-10624-y","volume":"57","author":"MM Saad","year":"2024","unstructured":"Saad MM, O\u2019Reilly R, Rehmani MH (2024) A survey on training challenges in generative adversarial networks for biomedical image analysis. Artificial Intelligence Review 57(2):19","journal-title":"Artificial Intelligence Review"},{"key":"20193_CR5","doi-asserted-by":"crossref","unstructured":"Yin H, Molchanov P, Alvarez JM, Li Z, Mallya A, Hoiem D, Jha NK, Kautz J (2020) Dreaming to distill: Data-free knowledge transfer via deepinversion, 2020. Paper presented at IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 14\u201319","DOI":"10.1109\/CVPR42600.2020.00874"},{"key":"20193_CR6","unstructured":"Fang G, Song J, Shen C, Wang X, Song M (2019) Data-free adversarial distillation, Preprint at https:\/\/arxiv.org\/abs\/1912.11006"},{"key":"20193_CR7","unstructured":"Luo L, Sandler M, Lin Z, Zhmoginov A, Howard A (2020) Large-scale generative data-free distillation, Preprint at https:\/\/arxiv.org\/abs\/2012.05578"},{"key":"20193_CR8","doi-asserted-by":"crossref","unstructured":"Vongkulbhisal J, Vinayavekhin P, Visentini-Scarzanella M (2019) Unifying heterogeneous classifiers with distillation, 2019. Paper presented at IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 175\u20133184","DOI":"10.1109\/CVPR.2019.00329"},{"issue":"1","key":"20193_CR9","first-page":"1","volume":"2022","author":"T Guiying","year":"2022","unstructured":"Guiying T (2022) Multimodal music emotion recognition method based on the combination of knowledge distillation and transfer learning. Sci Program 2022(1):1\u201313","journal-title":"Sci Program"},{"key":"20193_CR10","unstructured":"Hanting C, Tianyu G, Change X, Wenshuo L, Chunjing X, Chao X, Yunhe W (2021) Learning student networks in the wild, 2021. Paper presented at IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 6428\u20136437"},{"issue":"7","key":"20193_CR11","first-page":"8462","volume":"19","author":"T Wenxiao","year":"2022","unstructured":"Wenxiao T, Shakeel MS, Zisheng C, Hao W, Wenxiong K (2022) Target-category agnostic knowledge distillation with frequency domain supervision. IEEE Transact Ind Inform 19(7):8462\u20138471","journal-title":"IEEE Transact Ind Inform"},{"key":"20193_CR12","unstructured":"Mordvintsev A, Olah C, Tyka M (2015) Inceptionism:Going deeper into neural networks. https:\/\/research.googleblog.com\/2015\/06\/inceptionism-going-deeper-into-neural.html"},{"key":"20193_CR13","unstructured":"Bhardwaj K, Suda N, Marculescu R (2019) Dream distillation: A data-independent model compression framework, 2019. Paper presented at International Conference on Machine Learning (ICML), pp 1\u20134"},{"key":"20193_CR14","unstructured":"Nayak GK, Mopuri KR, Shaj V, Babu RV, Chakraborty A (2019) Zero-shot knowledge distillation in deep networks, 2019. Paper presented at International Conference on Machine Learning (ICML), pp 4743\u20134751"},{"key":"20193_CR15","unstructured":"Yuhang L, Feng Z, Ruihao G, Mingzhu S, Xin D, Fengwei Y, Shaoqing L, Shi G (2021) Mixmix: All you need for data-free compression are feature and data mixing, 2021. Paper presented at IEEE International Conference on Computer Vision (ICCV), pp 4410\u20134419"},{"key":"20193_CR16","unstructured":"Hanting C, Yunhe W, Chang X, Zhaohui Y, Chuanjian L, Boxin S, Chunjing X, Chao X, Qi T (2019) Data-free learning of student networks, 2019. Paper presented at IEEE International Conference on Computer Vision (ICCV), pp 3514\u20133522"},{"key":"20193_CR17","first-page":"9551","volume":"2","author":"P Micaelli","year":"2019","unstructured":"Micaelli P, Storkey A (2019) Zero-shot knowledge transfer via adversarial belief matching. Statistics 2:9551\u20139561","journal-title":"Statistics"},{"key":"20193_CR18","doi-asserted-by":"crossref","unstructured":"Pengchao H, Jihong P, Shiqing W, Yejun L (2021) Robustness and diversity seeking data-free knowledge distillation, 2021. Paper presented at IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 2740\u20132744","DOI":"10.1109\/ICASSP39728.2021.9414674"},{"key":"20193_CR19","doi-asserted-by":"publisher","first-page":"4262","DOI":"10.1109\/TMM.2022.3192663","volume":"24","author":"H Zhiwei","year":"2022","unstructured":"Zhiwei H, Yong L, Wang Z, Han H, Jianping A (2022) Cdfkd-mfs: Collaborative data-free knowledge distillation via multi-level feature sharing. IEEE Trans Multimedia 24:4262\u20134274","journal-title":"IEEE Trans Multimedia"},{"key":"20193_CR20","unstructured":"Gongfan F, Jie S, Xinchao W, Chenchao S, Xingen W, Mingji S (2021) Contrastive model inversion for data-free knowledge distillation, 2021. Paper presented at International Joint Conference on Artificial Intelligence (IJCAI), pp 1\u20137"},{"key":"20193_CR21","doi-asserted-by":"crossref","unstructured":"Jialiang T, Xiaoyan Y, Xin C, Ning J, Wenxin Y, Peng Z (2021) Data-free knowledge distillation with positive-unlabeled learning, 2021. Paper presented at International Conference on Neural Information Processing (ICONIP), pp 309\u2013320","DOI":"10.1007\/978-3-030-92270-2_27"},{"key":"20193_CR22","doi-asserted-by":"publisher","first-page":"116","DOI":"10.1109\/TIP.2022.3226416","volume":"32","author":"G Shiming","year":"2023","unstructured":"Shiming G, Bochao L, Pengju W, Yong L, Dan Z (2023) Learning privacy-preserving student networks via discriminative-generative distillation. IEEE Trans Image Process 32:116\u2013127","journal-title":"IEEE Trans Image Process"},{"key":"20193_CR23","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.neucom.2020.10.108","volume":"429","author":"Y Yongfei","year":"2021","unstructured":"Yongfei Y, Gehui S, Song Z, Ting H, Zhihong D, Unil Y (2021) Sequence generative adversarial nets with a conditional discriminator. Neurocomputing 429:69\u201376","journal-title":"Neurocomputing"},{"key":"20193_CR24","doi-asserted-by":"crossref","unstructured":"Mopuri KR, Uppala PK, Babu RV (2021) Ask, acquire, and attack: Data-free uap generative using class impressions, 2018. Paper presented at European Conference on Computer Vision(ECCV), pp 19\u201334","DOI":"10.1007\/978-3-030-01240-3_2"},{"key":"20193_CR25","doi-asserted-by":"crossref","unstructured":"Mahendran A, Vedaldi A (2015) Understanding deep image representations by inverting them, 2015. Paper presented at IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 5188\u20135196","DOI":"10.1109\/CVPR.2015.7299155"},{"issue":"9","key":"20193_CR26","doi-asserted-by":"publisher","first-page":"1066","DOI":"10.3390\/sym11091066","volume":"11","author":"M Kaya","year":"2019","unstructured":"Kaya M, Bilge HS (2019) Deep metric learning: A survey(review). Symmetry 11(9):1066\u20131092","journal-title":"Symmetry"},{"key":"20193_CR27","doi-asserted-by":"crossref","unstructured":"Kaiming H, Xoangyu Z, Shaoqing R, Jian S (2016) Deep residual learning for image recognition, 2016. Paper presented at IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"20193_CR28","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition, 2015. Paper presented at International Conference on Learning Representations, pp 1\u201314"},{"key":"20193_CR29","doi-asserted-by":"crossref","unstructured":"Zagoruyko S, Komodakis N (2016) Wide residual networks. Preprint at https:\/\/arxiv.org\/abs\/1605.07146","DOI":"10.5244\/C.30.87"},{"key":"20193_CR30","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium, 2017. Paper presented at International Conference on Neural Information Processing Systems (NeurIPS), pp 6629\u20136640"},{"key":"20193_CR31","doi-asserted-by":"crossref","unstructured":"Zhang R, Isola P, Efros AA, Shechtman E, Wang O (2018) The unreasonable effectiveness of deep features as a perceptual metric, 2018. Paper presented at IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 586\u2013595","DOI":"10.1109\/CVPR.2018.00068"},{"key":"20193_CR32","doi-asserted-by":"publisher","first-page":"104627","DOI":"10.1016\/j.imavis.2023.104627","volume":"131","author":"X Yu","year":"2023","unstructured":"Yu X, Yan L, Yang Y, Zhou L, Ou L (2023) Conditional generative data-free knowledge distillation. Image Vis Comput 131:104627. https:\/\/doi.org\/10.1016\/j.imavis.2023.104627. (ISSN 0262-8856)","journal-title":"Image Vis Comput"},{"issue":"10","key":"20193_CR33","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TPAMI.2023.3271626","volume":"45","author":"X Li","year":"2023","unstructured":"Li X, Wang S, Sun J, Zongben Xu (2023) Variational Data-Free Knowledge Distillation for Continual Learning. IEEE Trans Pattern Anal Mach Intell 45(10):1\u201317","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"13","key":"20193_CR34","first-page":"14991","volume":"38","author":"H Shin","year":"2024","unstructured":"Shin H, Choi DW (2024) Teacher as a Lenient Expert: Teacher-Agnostic Data-Free Knowledge Distillation. Proc AAAI Conf Artif Intell 38(13):14991\u201314999","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"20193_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TMM.2023.3236212","volume":"25","author":"X Li","year":"2023","unstructured":"Li X, Sun Q, Jiao L, Liu F, Liu X, Li L, Chen P, Zuo Y (2023) D3K: Dynastic Data-Free Knowledge Distillation. IEEE Transact Multimed 25:1\u201314","journal-title":"IEEE Transact Multimed"},{"key":"20193_CR36","doi-asserted-by":"crossref","unstructured":"Tran MT, Le T, Le XM et al (2024) Nayer: Noisy layer data generation for efficient and effective data-free knowledge distillation, 2024. Paper presented at IEEE\/CVF Conference on Computer Vision and Pattern Recogniton, pp 23860\u201323869","DOI":"10.1109\/CVPR52733.2024.02252"},{"issue":"2","key":"20193_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3639369","volume":"15","author":"Z Wu","year":"2024","unstructured":"Wu Z, Sun S, Wang Y et al (2024) Exploring the distributed knowledge congruence in proxy-data-free federated distillation. ACM Transact Intell Syst Technol 15(2):1\u201334","journal-title":"ACM Transact Intell Syst Technol"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-20193-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-20193-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-20193-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,5]],"date-time":"2025-09-05T22:30:23Z","timestamp":1757111423000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-20193-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,14]]},"references-count":37,"journal-issue":{"issue":"23","published-online":{"date-parts":[[2025,7]]}},"alternative-id":["20193"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-20193-z","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,9,14]]},"assertion":[{"value":"10 July 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 August 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 August 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 September 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of the article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}