{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T22:11:05Z","timestamp":1769897465234,"version":"3.49.0"},"publisher-location":"Cham","reference-count":19,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032145338","type":"print"},{"value":"9783032145345","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-14534-5_40","type":"book-chapter","created":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T09:23:46Z","timestamp":1769851426000},"page":"494-504","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Assessing Model Fusion Efficacy Through Hierarchical Knowledge Distillation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-4365-3955","authenticated-orcid":false,"given":"Bharat","family":"Choudhary","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0398-395X","authenticated-orcid":false,"given":"Deepak","family":"Singh","sequence":"additional","affiliation":[]},{"given":"Dilip","family":"Singh Sisodia","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,1]]},"reference":[{"key":"40_CR1","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network (2015). http:\/\/arxiv.org\/abs\/1503.02531. Accessed 09 Mar 2015"},{"key":"40_CR2","unstructured":"Zhang, L., Cao, Y., Chen, Y., Zhao, Y.: BERT model compression with decoupled knowledge distillation and representation learning. In: AISS 2022, pp. 25\u201327 (2022)"},{"key":"40_CR3","unstructured":"Cheng, Y., Wang, D., Zhou, P., Zhang, T.: A survey of model compression and acceleration for deep neural networks. IEEE Sign. Process. Mag., Spec. Issue Deep Learn. Image Underst. (2020)"},{"key":"40_CR4","doi-asserted-by":"publisher","first-page":"5113","DOI":"10.1007\/s10462-020-09816-7","volume":"53","author":"T Choudhary","year":"2020","unstructured":"Choudhary, T., Mishra, V., Goswami, A., Sarangapani, J.: A comprehensive survey on model compression and acceleration. Artif. Intell. Rev. 53, 5113\u20135155 (2020)","journal-title":"Artif. Intell. Rev."},{"issue":"6","key":"40_CR5","doi-asserted-by":"publisher","first-page":"1789","DOI":"10.1007\/s11263-021-01453-z","volume":"129","author":"J Gou","year":"2021","unstructured":"Gou, J., Baosheng, Yu., Maybank, S.J., Tao, D.: Knowledge distillation: a survey. Int. J. Comput. Vision 129(6), 1789\u20131819 (2021)","journal-title":"Int. J. Comput. Vision"},{"issue":"8","key":"40_CR6","doi-asserted-by":"publisher","first-page":"863","DOI":"10.1007\/s11265-020-01596-1","volume":"93","author":"A Berthelier","year":"2021","unstructured":"Berthelier, A., Chateau, T., Duffner, S., Garcia, C., Blanc, C.: Deep model compression and architecture optimization for embedded systems: a survey. J. Sign. Process. Syst. 93(8), 863\u2013878 (2021)","journal-title":"J. Sign. Process. Syst."},{"key":"40_CR7","unstructured":"Chowdhury, S., Liang, B., Tizghadam, A., Albanese, I.: Improving knowledge distillation with teacher\u2019s explanation. arXiv preprint: arXiv:2310.02572 (2023)"},{"key":"40_CR8","doi-asserted-by":"crossref","unstructured":"Ofori, M., El-Gayar, O., O\u2019Brien, A., Noteboom, C.: A deep learning model compression and ensemble approach for weed detection (2022)","DOI":"10.24251\/HICSS.2022.138"},{"issue":"13","key":"40_CR9","doi-asserted-by":"publisher","first-page":"4525","DOI":"10.1080\/00207543.2022.2160501","volume":"61","author":"Y Li","year":"2023","unstructured":"Li, Y., et al.: A hybrid model compression approach via knowledge distillation for predicting energy consumption in additive manufacturing. Int. J. Prod. Res. 61(13), 4525\u20134547 (2023)","journal-title":"Int. J. Prod. Res."},{"key":"40_CR10","doi-asserted-by":"publisher","first-page":"106","DOI":"10.1016\/j.neucom.2020.07.048","volume":"415","author":"Y Liu","year":"2020","unstructured":"Liu, Y., Zhang, W., Wang, J.: Adaptive multi-teacher multi-level knowledge distillation. Neurocomputing 415, 106\u2013113 (2020)","journal-title":"Neurocomputing"},{"key":"40_CR11","doi-asserted-by":"crossref","unstructured":"Panchbhai, K.G., Lanjewar, M.G., Malik, V.V., Charanarur, P.: Small size CNN (CAS-CNN), and modified MobileNetV2 (CAS-MODMOBNET) to identify cashew nut and fruit diseases. Multimedia Tools Appl. (2024)","DOI":"10.1007\/s11042-024-19042-w"},{"key":"40_CR12","doi-asserted-by":"crossref","unstructured":"Gundu, S.R., Panem, C.A., Thimmapuram, A., Gad, R.S.: Emerging computational challenges in cloud computing and RTEAH algorithm based solution. J. Ambient Intell. Human. Comput. (2022)","DOI":"10.1007\/s12652-021-03380-w"},{"key":"40_CR13","doi-asserted-by":"crossref","unstructured":"Lanjewar, M.G., Panchbhai, K.G., Charanarur, P.: Small size CNN-Based COVID-19 disease prediction system using CT scan images on PaaS cloud. Multimedia Tools Appl. 83(21) (2024)","DOI":"10.1007\/s11042-023-17884-4"},{"key":"40_CR14","doi-asserted-by":"crossref","unstructured":"Luo, S., Chen, D., Wang, C.: Knowledge distillation with deep supervision. In: 2023 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20138. IEEE (2023)","DOI":"10.1109\/IJCNN54540.2023.10191309"},{"key":"40_CR15","doi-asserted-by":"crossref","unstructured":"Charanarur, P., Thanh Hung, B., Chakrabarti, P., Siva Shankar, S.: Design optimization-based software-defined networking scheme for detecting and preventing attacks. Multimedia Tools Appl., 1\u201319 (2024)","DOI":"10.1007\/s11042-024-18466-8"},{"key":"40_CR16","doi-asserted-by":"crossref","unstructured":"Pang, Y., Zhang, Y., Wang, Y., Wei, X., Chen, B.: Exploring model compression limits and laws: a pyramid knowledge distillation framework for satellite-on-orbit object recognition. IEEE Trans. Geosci. Remote Sens. (2024)","DOI":"10.1109\/TGRS.2023.3348470"},{"key":"40_CR17","doi-asserted-by":"crossref","unstructured":"Mirzadeh, S.I., Farajtabar, M., Li, A., Levine, N., Matsukawa, A., Ghasemzadeh, H.: Improved knowledge distillation via teacher assistant. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, pp. 5191\u20135198 (2020)","DOI":"10.1609\/aaai.v34i04.5963"},{"key":"40_CR18","unstructured":"Harutyunyan, H., Rawat, A. S., Menon, A. K., Kim, S., Kumar, S.: Supervision complexity and its role in knowledge distillation. arXiv preprint: arXiv:2301.12245 (2023)"},{"key":"40_CR19","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1016\/j.neucom.2021.04.102","volume":"454","author":"T Wen","year":"2021","unstructured":"Wen, T., Lai, S., Qian, X.: Preparing lessons: improve knowledge distillation with better supervision. Neurocomputing 454, 25\u201333 (2021)","journal-title":"Neurocomputing"}],"container-title":["Communications in Computer and Information Science","Machine Learning, Image Processing, Network Security and Data Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-14534-5_40","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T09:23:48Z","timestamp":1769851428000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-14534-5_40"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9783032145338","9783032145345"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-14534-5_40","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"1 February 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MIND","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Machine Learning, Image Processing, Network Security and Data Sciences","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Goa","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mind2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/mind2024.mind-society.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}