{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T17:23:10Z","timestamp":1770830590632,"version":"3.50.1"},"reference-count":71,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"3","license":[{"start":{"date-parts":[[2023,9,1]],"date-time":"2023-09-01T00:00:00Z","timestamp":1693526400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,9,1]],"date-time":"2023-09-01T00:00:00Z","timestamp":1693526400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,9,1]],"date-time":"2023-09-01T00:00:00Z","timestamp":1693526400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61976107"],"award-info":[{"award-number":["61976107"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61502208"],"award-info":[{"award-number":["61502208"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Qing Lan Project of Colleges and Universities of Jiangsu Province in 2020"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Cogn. Dev. Syst."],"published-print":{"date-parts":[[2023,9]]},"DOI":"10.1109\/tcds.2022.3232569","type":"journal-article","created":{"date-parts":[[2022,12,28]],"date-time":"2022-12-28T18:47:17Z","timestamp":1672253237000},"page":"1574-1585","source":"Crossref","is-referenced-by-count":8,"title":["Channel-Correlation-Based Selective Knowledge Distillation"],"prefix":"10.1109","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1413-0693","authenticated-orcid":false,"given":"Jianping","family":"Gou","sequence":"first","affiliation":[{"name":"College of Computer and Information Science, College of Software, Southwest University, Chongqing, China"}]},{"given":"Xiangshuo","family":"Xiong","sequence":"additional","affiliation":[{"name":"School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, China"}]},{"given":"Baosheng","family":"Yu","sequence":"additional","affiliation":[{"name":"School of Computer Science, The University of Sydney, Darlington, NSW, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3180-0484","authenticated-orcid":false,"given":"Yibing","family":"Zhan","sequence":"additional","affiliation":[{"name":"JD Explore Academy, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5867-9322","authenticated-orcid":false,"given":"Zhang","family":"Yi","sequence":"additional","affiliation":[{"name":"School of Computer Science, Sichuan University, Chengdu, China"}]}],"member":"263","reference":[{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00409"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00511"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i8.16865"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00726"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2021\/362"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58610-2_2"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01163"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01241"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2019.2962094"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2020.3039952"},{"key":"ref11","article-title":"Distilling the knowledge in a neural network","author":"hinton","year":"2015","journal-title":"ArXiv 1503 02531"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00302"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3055564"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2021.3131045"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01252-6_17"},{"key":"ref16","article-title":"Uncertainty-aware contrastive distillation for incremental semantic segmentation","author":"yang","year":"2022","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17234"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01103"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1145\/3568679"},{"key":"ref50","article-title":"Channel-wise distillation for semantic segmentation","author":"shu","year":"2021","journal-title":"arXiv 2011 13256"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3212733"},{"key":"ref45","article-title":"Understanding and improving knowledge distillation","author":"tang","year":"2020","journal-title":"arXiv 2002 03532"},{"key":"ref48","article-title":"Channel distillation: Channel-wise attention for knowledge distillation","author":"zhao","year":"2020","journal-title":"arXiv 2006 01683"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref42","article-title":"Does knowledge distillation really work?","author":"stanton","year":"2021","journal-title":"arXiv 2106 05945"},{"key":"ref41","article-title":"Not all knowledge is created equal: Mutual distillation of confident knowledge","author":"li","year":"2021","journal-title":"arXiv 2106 01489"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/367"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00489"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-60636-7_21"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2017.2765695"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2020.3040796"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-021-01453-z"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2021.3123979"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2017.10.013"},{"key":"ref6","article-title":"A survey on green deep learning","author":"xu","year":"2021","journal-title":"arXiv 2111 05193"},{"key":"ref5","first-page":"9694","article-title":"Align before fuse: Vision and language representation learning with momentum distillation","author":"li","year":"2021","journal-title":"Advances in neural information processing systems"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20083-0_27"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3141255"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-82136-4_29"},{"key":"ref37","first-page":"22158","article-title":"Comprehensive knowledge distillation with causal intervention","author":"deng","year":"2021","journal-title":"Advances in neural information processing systems"},{"key":"ref36","article-title":"Collaborative teacher&#x2013;student learning via multiple knowledge transfer","author":"sun","year":"2021","journal-title":"arXiv 2101 08471"},{"key":"ref31","first-page":"1","article-title":"FitNets: Hints for thin deep nets","author":"romero","year":"2015","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref30","first-page":"13292","article-title":"Learning student-friendly teacher networks for knowledge distillation","author":"park","year":"2021","journal-title":"Advances in neural information processing systems"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01165"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00454"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2020.2965166"},{"key":"ref1","article-title":"Efficient visual recognition with deep neural networks: A survey on recent advances and new directions","author":"wu","year":"2021","journal-title":"arXiv 2108 13055"},{"key":"ref39","article-title":"Elastic knowledge distillation by learning from recollection","author":"fu","year":"2021","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01603"},{"key":"ref71","author":"guo","year":"2022","journal-title":"Reducing the teacher&#x2013;student gap via adaptive temperatures"},{"key":"ref70","first-page":"554","article-title":"3D object representations for fine-grained categorization","author":"krausel","year":"2013","journal-title":"Proc IEEE\/CVF Int Conf Comput Vis"},{"key":"ref24","first-page":"5316","article-title":"Even your teacher needs guidance: Ground-truth targets dampen regularization imposed by self-distillation","author":"borup","year":"2021","journal-title":"Advances in neural information processing systems"},{"key":"ref68","article-title":"Learning multiple layers of features from tiny images","author":"krizhevsky","year":"2009"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00396"},{"key":"ref67","article-title":"Knowledge distillation from a stronger teacher","author":"huang","year":"2022","journal-title":"arXiv 2205 10536"},{"key":"ref26","first-page":"7517","article-title":"Knowledge distillation by on-the-fly native ensemble","author":"zhu","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5963"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01389"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i2.16213"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00775"},{"key":"ref22","article-title":"Quantized reservoir computing for spectrum sensing with knowledge distillation","author":"liu","year":"2022","journal-title":"IEEE Trans Cogn Devel Syst"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00396"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3152732"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2021.3083113"},{"key":"ref28","article-title":"Cross-architecture knowledge distillation","author":"liu","year":"2022","journal-title":"arXiv 2207 05273"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2019.10.074"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58555-6_18"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6715"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00914"},{"key":"ref61","article-title":"Multi-level attention-based sample correlations for knowledge distillation","author":"gou","year":"2022","journal-title":"IEEE Trans Ind Informat"}],"container-title":["IEEE Transactions on Cognitive and Developmental Systems"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7274989\/10242607\/10000398.pdf?arnumber=10000398","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,2]],"date-time":"2023-10-02T18:08:31Z","timestamp":1696270111000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10000398\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9]]},"references-count":71,"journal-issue":{"issue":"3"},"URL":"https:\/\/doi.org\/10.1109\/tcds.2022.3232569","relation":{},"ISSN":["2379-8920","2379-8939"],"issn-type":[{"value":"2379-8920","type":"print"},{"value":"2379-8939","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,9]]}}}