{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T19:08:08Z","timestamp":1771614488539,"version":"3.50.1"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,5,4]],"date-time":"2022-05-04T00:00:00Z","timestamp":1651622400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"},{"start":{"date-parts":[[2022,5,4]],"date-time":"2022-05-04T00:00:00Z","timestamp":1651622400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002865","name":"Chongqing Science and Technology Commission","doi-asserted-by":"publisher","award":["cstc2020jscx-msxmX0086"],"award-info":[{"award-number":["cstc2020jscx-msxmX0086"]}],"id":[{"id":"10.13039\/501100002865","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007957","name":"Chongqing Municipal Education Commission","doi-asserted-by":"publisher","award":["KJQN202001137"],"award-info":[{"award-number":["KJQN202001137"]}],"id":[{"id":"10.13039\/501100007957","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002865","name":"Chongqing Science and Technology Commission","doi-asserted-by":"publisher","award":["cstc2019jscx-zdztzx0043"],"award-info":[{"award-number":["cstc2019jscx-zdztzx0043"]}],"id":[{"id":"10.13039\/501100002865","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002865","name":"Chongqing Science and Technology Commission","doi-asserted-by":"publisher","award":["cstc2019jcyj-msxmX0442"],"award-info":[{"award-number":["cstc2019jcyj-msxmX0442"]}],"id":[{"id":"10.13039\/501100002865","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,1]]},"DOI":"10.1007\/s10489-022-03486-4","type":"journal-article","created":{"date-parts":[[2022,5,9]],"date-time":"2022-05-09T16:03:12Z","timestamp":1652112192000},"page":"1997-2009","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":59,"title":["Teacher-student collaborative knowledge distillation for image classification"],"prefix":"10.1007","volume":"53","author":[{"given":"Chuanyun","family":"Xu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1609-7658","authenticated-orcid":false,"given":"Wenjian","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Tian","family":"Li","sequence":"additional","affiliation":[]},{"given":"Nanlan","family":"Bai","sequence":"additional","affiliation":[]},{"given":"Gang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,5,4]]},"reference":[{"key":"3486_CR1","first-page":"1097","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097\u20131105","journal-title":"Adv Neural Inf Process Syst"},{"key":"3486_CR2","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"3486_CR3","unstructured":"Tan M, Le Q (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning. PMLR, pp 6105\u20136114"},{"issue":"1","key":"3486_CR4","doi-asserted-by":"publisher","first-page":"126","DOI":"10.1109\/MSP.2017.2765695","volume":"35","author":"Y Cheng","year":"2018","unstructured":"Cheng Y, Wang D, Zhou P, Zhang T (2018) Model compression and acceleration for deep neural networks: The principles, progress, and challenges. IEEE Signal Proc Mag 35(1):126\u2013136","journal-title":"IEEE Signal Proc Mag"},{"key":"3486_CR5","doi-asserted-by":"crossref","unstructured":"Bashir D, Monta\u00f1ez GD, Sehra S, Segura PS, Lauw J (2020) An information-theoretic perspective on overfitting and underfitting. In: Australasian Joint Conference on Artificial Intelligence. Springer, pp 347\u2013358","DOI":"10.1007\/978-3-030-64984-5_27"},{"key":"3486_CR6","doi-asserted-by":"crossref","unstructured":"Yim J, Joo D, Bae J, Kim J (2017) A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4133\u20134141","DOI":"10.1109\/CVPR.2017.754"},{"key":"3486_CR7","doi-asserted-by":"crossref","unstructured":"Kim Y, Rush AM (2016) Sequence-level knowledge distillation. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp 1317\u20131327","DOI":"10.18653\/v1\/D16-1139"},{"issue":"6","key":"3486_CR8","doi-asserted-by":"publisher","first-page":"1789","DOI":"10.1007\/s11263-021-01453-z","volume":"129","author":"J Gou","year":"2021","unstructured":"Gou J, Yu B, Maybank SJ, Tao D (2021) Knowledge distillation: A survey. Int J Comput Vis 129(6):1789\u20131819","journal-title":"Int J Comput Vis"},{"key":"3486_CR9","doi-asserted-by":"crossref","unstructured":"Bucilua C, Caruana R, Niculescu-Mizil A (2006) Model compression. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp 535\u2013541","DOI":"10.1145\/1150402.1150464"},{"key":"3486_CR10","unstructured":"Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network"},{"key":"3486_CR11","unstructured":"Romero A, Ballas N, Kahou SE, Chassang A, Gatta C, Bengio Y (2015) Fitnets: Hints for thin deep nets. ICLR"},{"key":"3486_CR12","unstructured":"Komodakis N, Zagoruyko S (2017) Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. In: International Conference on Learning Representations"},{"key":"3486_CR13","doi-asserted-by":"crossref","unstructured":"Tung F, Mori G (2019) Similarity-preserving knowledge distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 1365\u20131374","DOI":"10.1109\/ICCV.2019.00145"},{"key":"3486_CR14","doi-asserted-by":"crossref","unstructured":"Ahn S, Hu SX, Damianou A, Lawrence ND, Dai Z (2019) Variational information distillation for knowledge transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 9163\u20139171","DOI":"10.1109\/CVPR.2019.00938"},{"key":"3486_CR15","doi-asserted-by":"crossref","unstructured":"Park W, Kim D, Lu Y, Cho M (2019) Relational knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 3967\u20133976","DOI":"10.1109\/CVPR.2019.00409"},{"key":"3486_CR16","unstructured":"Tian Y, Krishnan D, Isola P (2019) Contrastive representation distillation. In: International Conference on Learning Representations"},{"key":"3486_CR17","doi-asserted-by":"crossref","unstructured":"Heo B, Lee M, Yun S, Choi JY (2019) Knowledge transfer via distillation of activation boundaries formed by hidden neurons. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 33, pp 3779\u20133787","DOI":"10.1609\/aaai.v33i01.33013779"},{"key":"3486_CR18","doi-asserted-by":"crossref","unstructured":"Dong N, Zhang Y, Ding M, Xu S, Bai Y (2021) One-stage object detection knowledge distillation via adversarial learning. Appl Intell:1\u201317","DOI":"10.1007\/s10489-021-02634-6"},{"issue":"2","key":"3486_CR19","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1007\/s10489-020-01858-2","volume":"51","author":"OK Oyedotun","year":"2021","unstructured":"Oyedotun OK, Shabayek AER, Aouada D, Ottersten B (2021) Deep network compression with teacher latent subspace learning and lasso. Appl Intell 51(2):834\u2013853","journal-title":"Appl Intell"},{"key":"3486_CR20","unstructured":"Furlanello T, Lipton Z, Tschannen M, Itti L, Anandkumar A (2018) Born again neural networks. In: International Conference on Machine Learning. PMLR, pp 1607\u20131616"},{"key":"3486_CR21","doi-asserted-by":"crossref","unstructured":"Yuan L, Tay Francis EH, Li G, Wang T, Feng J (2020) Revisiting knowledge distillation via label smoothing regularization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 3903\u20133911","DOI":"10.1109\/CVPR42600.2020.00396"},{"key":"3486_CR22","unstructured":"Mobahi H, Farajtabar M, Bartlett P (2020) Self-distillation amplifies regularization in hilbert space. Neural Information Processing Systems (NeurIPS). https:\/\/papers.nips.cc\/paper\/2020\/file\/2288f691b58edecadcc9a8691762b4fd-Paper.pdf"},{"key":"3486_CR23","doi-asserted-by":"crossref","unstructured":"Phuong M, Lampert C H (2019) Distillation-based training for multi-exit architectures. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 1355\u20131364","DOI":"10.1109\/ICCV.2019.00144"},{"key":"3486_CR24","doi-asserted-by":"crossref","unstructured":"Zhang L, Song J, Gao A, Chen J, Bao C, Ma K (2019) Be your own teacher: Improve the performance of convolutional neural networks via self distillation. In: Proceedings of the IEEE\/ CVF International Conference on Computer Vision, pp 3713\u20133722","DOI":"10.1109\/ICCV.2019.00381"},{"key":"3486_CR25","doi-asserted-by":"crossref","unstructured":"Ji M, Shin S, Hwang S, Park G, Moon I-C (2021) Refine myself by teaching myself: Feature refinement via self-knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 10664\u2013 10673","DOI":"10.1109\/CVPR46437.2021.01052"},{"key":"3486_CR26","doi-asserted-by":"crossref","unstructured":"Yun S, Park J, Lee K, Shin J (2020) Regularizing class-wise predictions via self-knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 13876\u201313885","DOI":"10.1109\/CVPR42600.2020.01389"},{"key":"3486_CR27","doi-asserted-by":"crossref","unstructured":"Dietterich TG (2000) Ensemble methods in machine learning. In: International workshop on multiple classifier systems. Springer, pp 1\u201315","DOI":"10.1007\/3-540-45014-9_1"},{"issue":"1-2","key":"3486_CR28","doi-asserted-by":"publisher","first-page":"239","DOI":"10.1016\/S0004-3702(02)00190-X","volume":"137","author":"Z-H Zhou","year":"2002","unstructured":"Zhou Z-H, Wu J, Tang W (2002) Ensembling neural networks: many could be better than all. Artif Intell 137(1-2):239\u2013263","journal-title":"Artif Intell"},{"issue":"1","key":"3486_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10462-009-9124-7","volume":"33","author":"L Rokach","year":"2010","unstructured":"Rokach L (2010) Ensemble-based classifiers. Artif Intell Rev 33(1):1\u201339","journal-title":"Artif Intell Rev"},{"key":"3486_CR30","doi-asserted-by":"crossref","unstructured":"Fukuda T, Suzuki M, Kurata G, Thomas S, Cui J, Ramabhadran B (2017) Efficient knowledge distillation from an ensemble of teachers.. In: Interspeech, pp 3697\u20133701","DOI":"10.21437\/Interspeech.2017-614"},{"key":"3486_CR31","unstructured":"Lan X, Zhu X, Gong S (2018) Knowledge distillation by on-the-fly native ensemble. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp 7528\u20137538"},{"key":"3486_CR32","doi-asserted-by":"publisher","first-page":"106","DOI":"10.1016\/j.neucom.2020.07.048","volume":"415","author":"Y Liu","year":"2020","unstructured":"Liu Y, Zhang W, Wang J (2020) Adaptive multi-teacher multi-level knowledge distillation. Neurocomputing 415:106\u2013113","journal-title":"Neurocomputing"},{"key":"3486_CR33","unstructured":"Krizhevsky A (2009) Learning multiple layers of features from tiny images. Master\u2019s thesis, University of Tront"},{"key":"3486_CR34","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"3486_CR35","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Bengio Y, LeCun Y (eds) 3rd International Conference on Learning Representations, ICLR 2015, Conference Track Proceedings, San Diego"},{"key":"3486_CR36","doi-asserted-by":"crossref","unstructured":"Zagoruyko S, Komodakis N (2016) Wide residual networks. In: British Machine Vision Conference 2016. British Machine Vision Association","DOI":"10.5244\/C.30.87"},{"key":"3486_CR37","doi-asserted-by":"crossref","unstructured":"Zhang X, Zhou X, Lin M, Sun J (2018) Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848\u20136856","DOI":"10.1109\/CVPR.2018.00716"},{"key":"3486_CR38","doi-asserted-by":"crossref","unstructured":"Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: Practical guidelines for efficient cnn architecture design. In: Proceedings of the European conference on computer vision (ECCV), pp 116\u2013131","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"3486_CR39","doi-asserted-by":"crossref","unstructured":"Peng B, Jin X, Liu J, Li D, Wu Y, Liu Y, Zhou S, Zhang Z (2019) Correlation congruence for knowledge distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 5007\u20135016","DOI":"10.1109\/ICCV.2019.00511"},{"key":"3486_CR40","doi-asserted-by":"crossref","unstructured":"Passalis N, Tefas A (2018) Learning deep representations with probabilistic knowledge transfer. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 268\u2013284","DOI":"10.1007\/978-3-030-01252-6_17"},{"key":"3486_CR41","unstructured":"Kim J, Park S, Kwak N (2018) Paraphrasing complex network: network compression via factor transfer. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp 2765\u20132774"},{"key":"3486_CR42","doi-asserted-by":"crossref","unstructured":"Xu G, Liu Z, Li X, Loy CC (2020) Knowledge distillation meets self-supervision. In: European Conference on Computer Vision. Springer, pp 588\u2013604","DOI":"10.1007\/978-3-030-58545-7_34"},{"key":"3486_CR43","doi-asserted-by":"crossref","unstructured":"Yang C, An Z, Cai L, Xu Y (2021) Hierarchical self-supervised augmented knowledge distillation. In: Zhou Z-H (ed) Proceedings of the thirtieth international joint conference on artificial intelligence, IJCAI-21, pp 1217\u20131223","DOI":"10.24963\/ijcai.2021\/168"},{"key":"3486_CR44","doi-asserted-by":"crossref","unstructured":"Ji M, Heo B, Park S (2021) Show, attend and distill: Knowledge distillation via attention-based feature matching. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 35, pp 7945\u20137952","DOI":"10.1609\/aaai.v35i9.16969"},{"key":"3486_CR45","unstructured":"Lee C-Y, Xie S, Gallagher P, Zhang Z, Tu Z (2015) Deeply-supervised nets. In: Artificial intelligence and statistics. PMLR, pp 562\u2013570"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03486-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-022-03486-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03486-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,4]],"date-time":"2023-01-04T05:04:48Z","timestamp":1672808688000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-022-03486-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,4]]},"references-count":45,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,1]]}},"alternative-id":["3486"],"URL":"https:\/\/doi.org\/10.1007\/s10489-022-03486-4","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,5,4]]},"assertion":[{"value":"7 March 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 May 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}