{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T15:14:43Z","timestamp":1769181283846,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":16,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556816","type":"print"},{"value":"9789819556823","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5682-3_10","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:14:33Z","timestamp":1769116473000},"page":"136-149","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Self-distillation with\u00a0Mutual Assistance Mechanism: Enhancing Model Performance Through Collaborative Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-8528-4618","authenticated-orcid":false,"given":"Lin","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2779-582X","authenticated-orcid":false,"given":"Zhe","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-5838-7685","authenticated-orcid":false,"given":"Yuqing","family":"Tian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"key":"10_CR1","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv:1503.02531 (2015)"},{"issue":"11","key":"10_CR2","doi-asserted-by":"publisher","first-page":"10475","DOI":"10.1109\/TMC.2024.3374706","volume":"23","author":"L Yi","year":"2024","unstructured":"Yi, L., Shi, X., Wang, N., Zhang, J., Wang, G., Liu, X.: Fedpe: adaptive model pruning-expanding for federated learning on mobile devices. IEEE Trans. Mob. Comput. 23(11), 10475\u201310493 (2024)","journal-title":"IEEE Trans. Mob. Comput."},{"key":"10_CR3","doi-asserted-by":"publisher","first-page":"3265","DOI":"10.1109\/TIFS.2024.3360891","volume":"19","author":"Y Yang","year":"2024","unstructured":"Yang, Y., et al.: Quantization aware attack: enhancing transferable adversarial attacks by model quantization. IEEE Trans. Inf. Forensics Secur. 19, 3265\u20133278 (2024)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"10_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, L., Song, J., Gao, A., Chen, J., Bao, C., Ma, K.: Be your own teacher: improve the performance of convolutional neural networks via self distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3713\u20133722 (2019)","DOI":"10.1109\/ICCV.2019.00381"},{"key":"10_CR5","unstructured":"Romero, A., Ballas, N., Kahou, S.E., Chassang, A., Gatta, C., Bengio, Y.: Fitnets: hints for thin deep nets. arXiv:1412.6550 (2014)"},{"key":"10_CR6","doi-asserted-by":"crossref","unstructured":"Mirzadeh, S.I., Farajtabar, M., Li, A., Levine, N., Matsukawa, A., Ghasemzadeh, H.: Improved knowledge distillation via teacher assistant. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 5191\u20135198 (2020)","DOI":"10.1609\/aaai.v34i04.5963"},{"issue":"8","key":"10_CR7","first-page":"4388","volume":"44","author":"L Zhang","year":"2021","unstructured":"Zhang, L., Bao, C., Ma, K.: Self-distillation: towards efficient and compact neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 44(8), 4388\u20134403 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"11","key":"10_CR8","doi-asserted-by":"publisher","first-page":"8743","DOI":"10.1109\/TNNLS.2022.3152732","volume":"34","author":"S Li","year":"2022","unstructured":"Li, S., et al.: Distilling a powerful student model via online knowledge distillation. IEEE Trans. Neural Netw. Learn. Syst. 34(11), 8743\u20138752 (2022)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10_CR9","doi-asserted-by":"crossref","unstructured":"Yang, C., et al.: Mixskd: self-knowledge distillation from mixup for image recognition. In: European Conference on Computer Vision, pp. 534\u2013551 (2022)","DOI":"10.1007\/978-3-031-20053-3_31"},{"issue":"1","key":"10_CR10","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."},{"key":"10_CR11","doi-asserted-by":"crossref","unstructured":"Son, W., Na, J., Choi, J., Hwang, W.: Densely guided knowledge distillation using multiple teacher assistants. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9395\u20139404 (2021)","DOI":"10.1109\/ICCV48922.2021.00926"},{"key":"10_CR12","unstructured":"Zagoruyko, S., Komodakis, N.: Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. arXiv:1612.03928 (2016)"},{"key":"10_CR13","doi-asserted-by":"crossref","unstructured":"Tung, F., Mori, G.: Similarity-preserving knowledge distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1365\u20131374 (2019)","DOI":"10.1109\/ICCV.2019.00145"},{"key":"10_CR14","doi-asserted-by":"crossref","unstructured":"Yun, S., Park, J., Lee, K., Shin, J.: Regularizing class-wise predictions via self-knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13876\u201313885 (2020)","DOI":"10.1109\/CVPR42600.2020.01389"},{"key":"10_CR15","doi-asserted-by":"crossref","unstructured":"Hou, Y., Ma, Z., Liu, C., Loy, C.C.: Learning lightweight lane detection CNNs by self attention distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1013\u20131021 (2019)","DOI":"10.1109\/ICCV.2019.00110"},{"key":"10_CR16","doi-asserted-by":"crossref","unstructured":"Yuan, L., Tay, F.E., Li, G., Wang, T., Feng, J.: Revisiting knowledge distillation via label smoothing regularization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3903\u20133911 (2020)","DOI":"10.1109\/CVPR42600.2020.00396"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5682-3_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:14:36Z","timestamp":1769116476000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5682-3_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556816","9789819556823"],"references-count":16,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5682-3_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}