{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T05:35:07Z","timestamp":1730266507520,"version":"3.28.0"},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,7,18]]},"DOI":"10.1109\/ijcnn52387.2021.9534441","type":"proceedings-article","created":{"date-parts":[[2021,9,21]],"date-time":"2021-09-21T20:40:52Z","timestamp":1632256852000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["Triplet Knowledge Distillation Networks for Model Compression"],"prefix":"10.1109","author":[{"given":"Jialiang","family":"Tang","sequence":"first","affiliation":[]},{"given":"Ning","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Wenxin","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Wenqin","family":"Wu","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","first-page":"451","article-title":"Metric learning by collapsing classes","author":"globerson","year":"2005","journal-title":"Advances in Neural Information Processing Systems 18 [Neural Information Processing Systems NIPS 2005 December 5&#x2013;8 2005 Vancouver British Columbia Canada]"},{"key":"ref32","first-page":"1473","article-title":"Distance metric learning for large margin nearest neighbor classification","author":"weinberger","year":"2005","journal-title":"Advances in Neural Information Processing Systems 18 [Neural Information Processing Systems NIPS 2005 December 5&#x2013;8 2005 Vancouver British Columbia Canada]"},{"key":"ref31","article-title":"Training shallow and thin networks for acceleration via knowledge distillation with conditional adversarial networks","author":"xu","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-35289-8_25"},{"key":"ref34","first-page":"207","article-title":"Distance metric learning for large margin nearest neighbor classification","volume":"10","author":"weinberger","year":"2009","journal-title":"J Mach Learn Res"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00511"},{"journal-title":"Positive-unlabeled compression on the cloud","year":"2019","author":"xu","key":"ref11"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00361"},{"key":"ref13","article-title":"Layer-wise data-free CNN compression","author":"horton","year":"2020","journal-title":"CoRR vol abs\/2011 09058"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.298"},{"key":"ref15","first-page":"3299","article-title":"meprop: Sparsified back propagation for accelerated deep learning with reduced overfitting","author":"sun","year":"0","journal-title":"Proceedings of the 34th International Conference on Machine Learning-Volume 70"},{"key":"ref16","article-title":"Network pruning via transformable architecture search","author":"dong","year":"2019","journal-title":"CoRR vol abs\/1905 09717"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00447"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00290"},{"key":"ref19","article-title":"Rethinking the value of network pruning","author":"liu","year":"2019","journal-title":"7th International Conference on Learning Representations ICLR 2019"},{"key":"ref28","article-title":"Wide residual networks","author":"zagoruyko","year":"2016","journal-title":"Proceedings of the British Machine Vision Conference 2016 BMVC 2016 York UK September 19&#x2013;22 2016"},{"key":"ref4","article-title":"meprop: Sparsified back propagation for accelerated deep learning with reduced overfitting","author":"xu","year":"2017","journal-title":"ICML 2017"},{"key":"ref27","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2014","journal-title":"Computer Science"},{"key":"ref3","article-title":"Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer","author":"zagoruyko","year":"2016","journal-title":"ArXiv Preprint"},{"journal-title":"Reading digits in natural images with unsupervised feature learning","year":"2011","author":"netzer","key":"ref6"},{"key":"ref29","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2015","journal-title":"3rd International Conference on Learning Representations ICLR 2015"},{"key":"ref5","article-title":"The lottery ticket hypothesis: Finding sparse, trainable neural networks","author":"frankle","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref8","first-page":"215","article-title":"An analysis of single-layer networks in unsupervised feature learning","volume":"15","author":"coates","year":"2011","journal-title":"Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics AISTATS 2011"},{"journal-title":"Learning multiple layers of features from tiny images","year":"2009","author":"krizhevsky","key":"ref7"},{"key":"ref2","first-page":"38","article-title":"Distilling the knowledge in a neural network","volume":"14","author":"hinton","year":"2015","journal-title":"Computer Science"},{"key":"ref9","article-title":"Fitnets: Hints for thin deep nets","author":"romero","year":"2014","journal-title":"Computer Science"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref20","article-title":"Compressing deep convolutional networks using vector quantization","author":"gong","year":"2014","journal-title":"CoRR vol abs\/1412 6115"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.521"},{"journal-title":"Binarized neural networks Training deep neural networks with weights and activations constrained to + I or -1","year":"2016","author":"courbariaux","key":"ref21"},{"key":"ref24","first-page":"807","article-title":"Rectified linear units improve restricted boltzmann machines","author":"nair","year":"2010","journal-title":"Proceedings of the 27th International Conference on Machine Learning (ICML-10) June 21&#x2013;24 2010 Haifa Israel"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1142\/S0218001493000339"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24261-3_7"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2018.8451102"}],"event":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2021,7,18]]},"location":"Shenzhen, China","end":{"date-parts":[[2021,7,22]]}},"container-title":["2021 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9533266\/9533267\/09534441.pdf?arnumber=9534441","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:45:47Z","timestamp":1652197547000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9534441\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,18]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/ijcnn52387.2021.9534441","relation":{},"subject":[],"published":{"date-parts":[[2021,7,18]]}}}