{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T21:54:03Z","timestamp":1775598843556,"version":"3.50.1"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,26]],"date-time":"2024-09-26T00:00:00Z","timestamp":1727308800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,26]],"date-time":"2024-09-26T00:00:00Z","timestamp":1727308800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Anhui University of Science and Technology 2023 Graduate Innovation Fund Project","award":["2023CX2136"],"award-info":[{"award-number":["2023CX2136"]}]},{"name":"Medical Special Cultivation Project of Anhui University of Science and Technology","award":["YZ2023H2C005"],"award-info":[{"award-number":["YZ2023H2C005"]}]},{"name":"Research Foundation of the Institute of Environment-friendly Materials and Occupational Health (Wuhu) Anhui University of Science and Technology","award":["ALW2021YF04"],"award-info":[{"award-number":["ALW2021YF04"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s00530-024-01483-w","type":"journal-article","created":{"date-parts":[[2024,9,26]],"date-time":"2024-09-26T19:02:04Z","timestamp":1727377324000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["ATMKD: adaptive temperature guided multi-teacher knowledge distillation"],"prefix":"10.1007","volume":"30","author":[{"given":"Yu-e","family":"Lin","sequence":"first","affiliation":[]},{"given":"Shuting","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Yifeng","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Xingzhu","family":"Liang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,26]]},"reference":[{"key":"1483_CR1","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.-T., Sun, J.: Shufflenet v2: Practical guidelines for efficient cnn architecture design, In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 116\u2013131 (2018).","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"1483_CR2","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2117\u20132125 (2017).","DOI":"10.1109\/CVPR.2017.106"},{"key":"1483_CR3","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3431\u20133440 (2015).","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"1483_CR4","unstructured":"Hinton, G., Vinyals, O., Dean, J., Distilling the knowledge in a neural network, arXiv preprint arXiv:150302531, (2015)."},{"key":"1483_CR5","unstructured":"Adriana, R., Nicolas, B., Ebrahimi, K. S., Antoine, C., Carlo, G., Yoshua, B.: Fitnets: Hints for thin deep nets, In: Proceedings of the International Conference on Learning Representations (ICLR), pp. 1 (2015)."},{"key":"1483_CR6","unstructured":"Zagoruyko, S., Komodakis, N., Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer, arXiv preprint arXiv:161203928, (2016)."},{"key":"1483_CR7","doi-asserted-by":"crossref","unstructured":"Passalis, N., Tefas, A.: Learning deep representations with probabilistic knowledge transfer, In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 268\u2013284 (2018).","DOI":"10.1007\/978-3-030-01252-6_17"},{"key":"1483_CR8","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-024-03414-2","author":"G Yang","year":"2024","unstructured":"Yang, G., Ding, Y., Fang, X., Zhang, J., Chu, Y.: Fast face swapping with high-fidelity lightweight generator assisted by online knowledge distillation. Vis. Comput. (2024). https:\/\/doi.org\/10.1007\/s00371-024-03414-2","journal-title":"Vis. Comput."},{"key":"1483_CR9","doi-asserted-by":"crossref","unstructured":"Heo, B., Lee, M., Yun, S., Choi, J. Y.: Knowledge transfer via distillation of activation boundaries formed by hidden neurons, In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 3779\u20133787 (2019).","DOI":"10.1609\/aaai.v33i01.33013779"},{"key":"1483_CR10","unstructured":"Kim, J., Park, S., Kwak, N., Paraphrasing complex network: Network compression via factor transfer, Advances in Neural Information Processing Systems (NeurIPS), 31, (2018)."},{"key":"1483_CR11","doi-asserted-by":"crossref","unstructured":"You, S., Xu, C., Xu, C., Tao, D.: Learning from multiple teacher networks, In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pp. 1285\u20131294 (2017).","DOI":"10.1145\/3097983.3098135"},{"key":"1483_CR12","doi-asserted-by":"crossref","unstructured":"Fukuda, T., Suzuki, M., Kurata, G., Thomas, S., Cui, J., Ramabhadran, B.: Efficient Knowledge Distillation from an Ensemble of Teachers, In: Conference of the International Speech Communication Association (INTERSPEECH), pp. 3697\u20133701 (2017).","DOI":"10.21437\/Interspeech.2017-614"},{"key":"1483_CR13","doi-asserted-by":"crossref","unstructured":"Zhang, H., Chen, D., Wang, C.: Confidence-aware multi-teacher knowledge distillation, In: ICASSP 2022\u20132022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4498\u20134502 (2022).","DOI":"10.1109\/ICASSP43922.2022.9747534"},{"key":"1483_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, H., Chen, D., Wang, C.: Adaptive multi-teacher knowledge distillation with meta-learning, In: 2023 IEEE International Conference on Multimedia and Expo (ICME), pp. 1943\u20131948 (2023)..","DOI":"10.1109\/ICME55011.2023.00333"},{"key":"1483_CR15","doi-asserted-by":"crossref","unstructured":"Yuan, F., Shou, L., Pei, J., Lin, W., Gong, M., Fu, Y., Jiang, D.: Reinforced multi-teacher selection for knowledge distillation, In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 14284\u201314291 (2021).","DOI":"10.1609\/aaai.v35i16.17680"},{"key":"1483_CR16","doi-asserted-by":"crossref","unstructured":"Caubri\u00e8re, A., Tomashenko, N., Laurent, A., Morin, E., Camelin, N., Esteve, Y.: Curriculum-based transfer learning for an effective end-to-end spoken language understanding and domain portability, In: Conference of the International Speech Communication Association (INTERSPEECH), 2019).","DOI":"10.21437\/Interspeech.2019-1832"},{"key":"1483_CR17","doi-asserted-by":"crossref","unstructured":"Duan, Y., Zhu, H., Wang, H., Yi, L., Nevatia, R., Guibas, L. J.: Curriculum deepsdf, In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 51\u201367 (2020).","DOI":"10.1007\/978-3-030-58598-3_4"},{"key":"1483_CR18","doi-asserted-by":"crossref","unstructured":"Xiang, L., Ding, G., Han, J.: Learning from multiple experts: Self-paced knowledge distillation for long-tailed classification, In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 247\u2013263 (2020).","DOI":"10.1007\/978-3-030-58558-7_15"},{"key":"1483_CR19","doi-asserted-by":"crossref","unstructured":"Peng, B., Jin, X., Liu, J., Li, D., Wu, Y., Liu, Y., Zhou, S., Zhang, Z.: Correlation congruence for knowledge distillation, In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5007\u20135016 (2019).","DOI":"10.1109\/ICCV.2019.00511"},{"key":"1483_CR20","unstructured":"Liu, J. L. A. B. L. A. H. L. A. Y., Meta Knowledge Distillation, arXiv preprint arXiv:220207940, (2022)."},{"key":"1483_CR21","unstructured":"Ganin, Y., Lempitsky, V.: Unsupervised domain adaptation by backpropagation, In: International Conference on Machine Learning (ICML), pp. 1180\u20131189 (2015)."},{"key":"1483_CR22","doi-asserted-by":"crossref","unstructured":"Tung, F., Mori, G.: Similarity-preserving knowledge distillation, In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 1365\u20131374 (2019).","DOI":"10.1109\/ICCV.2019.00145"},{"key":"1483_CR23","doi-asserted-by":"crossref","unstructured":"Ahn, S., Hu, S. X., Damianou, A., Lawrence, N. D., Dai, Z.: Variational information distillation for knowledge transfer, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9163\u20139171 (2019).","DOI":"10.1109\/CVPR.2019.00938"},{"key":"1483_CR24","unstructured":"Tian, Y., Krishnan, D., Isola, P., Contrastive representation distillation, arXiv preprint arXiv:191010699, (2019)."},{"key":"1483_CR25","doi-asserted-by":"crossref","unstructured":"Yim, J., Joo, D., Bae, J., Kim, J.: A gift from knowledge distillation: Fast optimization, network minimization and transfer learning, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4133\u20134141 (2017).","DOI":"10.1109\/CVPR.2017.754"},{"key":"1483_CR26","doi-asserted-by":"crossref","unstructured":"Zhao, B., Cui, Q., Song, R., Qiu, Y., Liang, J.: Decoupled knowledge distillation, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11953\u201311962 (2022).","DOI":"10.1109\/CVPR52688.2022.01165"},{"key":"1483_CR27","doi-asserted-by":"crossref","unstructured":"Jin, Y., Wang, J., Lin, D.: Multi-level logit distillation, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 24276\u201324285 (2023).","DOI":"10.1109\/CVPR52729.2023.02325"},{"key":"1483_CR28","doi-asserted-by":"publisher","first-page":"106","DOI":"10.1016\/J.NEUCOM.2020.07.048","volume":"415","author":"Y Liu","year":"2020","unstructured":"Liu, Y., Zhang, W., Wang, J.: Adaptive multi-teacher multi-level knowledge distillation. Neurocomputing 415, 106\u2013113 (2020). https:\/\/doi.org\/10.1016\/J.NEUCOM.2020.07.048","journal-title":"Neurocomputing"},{"key":"1483_CR29","doi-asserted-by":"crossref","unstructured":"Kwon, K., Na, H., Lee, H., Kim, N. S.: Adaptive knowledge distillation based on entropy, In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 7409\u20137413 (2020).","DOI":"10.1109\/ICASSP40776.2020.9054698"},{"key":"1483_CR30","first-page":"12345","volume":"33","author":"S Du","year":"2020","unstructured":"Du, S., You, S., Li, X., Wu, J., Wang, F., Qian, C., Zhang, C.: Agree to disagree: Adaptive ensemble knowledge distillation in gradient space. Adv. Neural. Inf. Process. Syst. (NeurIPS) 33, 12345\u201312355 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst. (NeurIPS)"},{"key":"1483_CR31","doi-asserted-by":"crossref","unstructured":"Ding, Z., Jiang, G., Zhang, S., Guo, L., Lin, W.: How to Trade Off the Quantity and Capacity of Teacher Ensemble: Learning Categorical Distribution to Stochastically Employ a Teacher for Distillation, In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 17915\u201317923 (2024).","DOI":"10.1609\/aaai.v38i16.29746"},{"key":"1483_CR32","first-page":"21653","volume":"33","author":"S Sinha","year":"2020","unstructured":"Sinha, S., Garg, A., Larochelle, H.: Curriculum by smoothing. Adv. Neural. Inf. Process. Syst. (NeurIPS) 33, 21653\u201321664 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst. (NeurIPS)"},{"key":"1483_CR33","doi-asserted-by":"publisher","DOI":"10.18653\/V1\/P19-1486","author":"Y Tay","year":"2019","unstructured":"Tay, Y., Wang, S., Tuan, L.A., Fu, J., Phan, M.C., Yuan, X., Rao, J., Hui, S.C., Zhang, A.: Simple and effective curriculum pointer-generator networks for reading comprehension over long narratives. Assoc. Comput. Linguist. (ACL) (2019). https:\/\/doi.org\/10.18653\/V1\/P19-1486","journal-title":"Assoc. Comput. Linguist. (ACL)"},{"issue":"9","key":"1483_CR34","doi-asserted-by":"publisher","first-page":"4555","DOI":"10.1109\/TPAMI.2021.3069908","volume":"44","author":"X Wang","year":"2021","unstructured":"Wang, X., Chen, Y., Zhu, W.: A survey on curriculum learning. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 4555\u20134576 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2021.3069908","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1483_CR35","doi-asserted-by":"crossref","unstructured":"Yu, L., Weng, Z., Wang, Y., Zhu, Y.: Multi-teacher knowledge distillation for incremental implicitly-refined classification, In: IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136 (2022).","DOI":"10.1109\/ICME52920.2022.9859936"},{"key":"1483_CR36","doi-asserted-by":"crossref","unstructured":"Pham, C., Hoang, T., Do, T.-T.: Collaborative multi-teacher knowledge distillation for learning low bit-width deep neural networks, In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 6435\u20136443 (2023).","DOI":"10.1109\/WACV56688.2023.00637"},{"key":"1483_CR37","unstructured":"Guo, J., Reducing the teacher-student gap via adaptive temperatures, (2021)."},{"key":"1483_CR38","first-page":"3830","volume":"35","author":"X-C Li","year":"2022","unstructured":"Li, X.-C., Fan, W.-S., Song, S., Li, Y., Yunfeng, S., Zhan, D.-C.: Asymmetric temperature scaling makes larger networks teach well again. Adv. Neural Inf. Process. Syst. (NeurIPS) 35, 3830\u20133842 (2022)","journal-title":"Adv. Neural Inf. Process. Syst. (NeurIPS)"},{"key":"1483_CR39","unstructured":"Zheng, K., Yang, E.-H., Knowledge Distillation Based on Transformed Teacher Matching, arXiv preprint arXiv:240211148, (2024)."},{"key":"1483_CR40","doi-asserted-by":"crossref","unstructured":"Chen, D., Mei, J.-P., Zhang, H., Wang, C., feng, Y., Chen, C.: Knowledge distillation with the reused teacher classifier, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11933\u201311942 (2022).","DOI":"10.1109\/CVPR52688.2022.01163"},{"key":"1483_CR41","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y., Generative adversarial nets, Advances in Neural Information Processing Systems (NeurIPS), 27, (2014)."},{"key":"1483_CR42","unstructured":"Krizhevsky, A., Hinton, G., Learning multiple layers of features from tiny images, (2009)."},{"key":"1483_CR43","unstructured":"Le, Y., Yang, X., Tiny imagenet visual recognition challenge, CS 231N, 7 (7), 3 (2015)"},{"key":"1483_CR44","unstructured":"Khosla, A., Jayadevaprakash, N., Yao, B., Li, F.-F.: Novel dataset for fine-grained image categorization: Stanford dogs, In: Proc CVPR workshop on fine-grained visual categorization (FGVC), 2011)."},{"key":"1483_CR45","doi-asserted-by":"crossref","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition, In: Proceedings of the International Conference on Learning Representations (ICLR), 2015).","DOI":"10.1109\/ICCV.2015.314"},{"key":"1483_CR46","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016).","DOI":"10.1109\/CVPR.2016.90"},{"key":"1483_CR47","doi-asserted-by":"crossref","unstructured":"Zagoruyko, S., Komodakis, N.: Wide residual networks, In: Proceedings of the British Machine Vision Conference (BMVC), 2016).","DOI":"10.5244\/C.30.87"},{"key":"1483_CR48","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: Mobilenetv2: Inverted residuals and linear bottlenecks, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4510\u20134520 (2018).","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1483_CR49","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: An extremely efficient convolutional neural network for mobile devices, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6848\u20136856 (2018).","DOI":"10.1109\/CVPR.2018.00716"},{"key":"1483_CR50","doi-asserted-by":"crossref","unstructured":"Chen, D., Mei, J.-P., Zhang, Y., Wang, C., Wang, Z., Feng, Y., Chen, C.: Cross-layer distillation with semantic calibration, In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 7028\u20137036 (2021).","DOI":"10.1609\/aaai.v35i8.16865"},{"key":"1483_CR51","unstructured":"Yang, J., Martinez, B., Bulat, A., Tzimiropoulos, G.: Knowledge distillation via softmax regression representation learning, In: International Conference on Learning Representations (ICLR), 2020)"},{"key":"1483_CR52","doi-asserted-by":"crossref","unstructured":"Li, Z., Li, X., Yang, L., Zhao, B., Song, R., Luo, L., Li, J., Yang, J.: Curriculum temperature for knowledge distillation, In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 1504\u20131512 (2023).","DOI":"10.1609\/aaai.v37i2.25236"},{"key":"1483_CR53","unstructured":"Chi, Z., Zheng, T., Li, H., Yang, Z., Wu, B., Lin, B., Cai, D., Normkd: Normalized logits for knowledge distillation, arXiv preprint arXiv:230800520, (2023)."},{"key":"1483_CR54","doi-asserted-by":"crossref","unstructured":"Guo, Z., Yan, H., Li, H., Lin, X.: Class attention transfer based knowledge distillation, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11868\u201311877 (2023).","DOI":"10.1109\/CVPR52729.2023.01142"},{"key":"1483_CR55","doi-asserted-by":"crossref","unstructured":"Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: Visual explanations from deep networks via gradient-based localization, In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 618\u2013626 (2017).","DOI":"10.1109\/ICCV.2017.74"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01483-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01483-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01483-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T19:38:55Z","timestamp":1732822735000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01483-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,26]]},"references-count":55,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1483"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01483-w","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9,26]]},"assertion":[{"value":"29 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 September 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 September 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This study was conducted with the highest regard for ethical standards and following relevant guidelines and regulations. The research protocol did not require ethical review or approval as it did not involve human participants, animals, or sensitive data. All data used in this study were obtained from publicly available sources and were properly cited and acknowledged. No private or personally identifiable information was used or accessed during this research. The authors declare that there are no conflicts of interest, financial or otherwise, that could potentially influence the objectivity or integrity of this study. While no ethical review or approval was necessary for this particular study, the principles of academic integrity and research ethics were strictly adhered to throughout the research process.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"292"}}