{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T01:10:41Z","timestamp":1740100241696,"version":"3.37.3"},"reference-count":39,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62076084"],"award-info":[{"award-number":["62076084"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,7,18]]},"DOI":"10.1109\/ijcnn52387.2021.9533623","type":"proceedings-article","created":{"date-parts":[[2021,9,20]],"date-time":"2021-09-20T21:27:41Z","timestamp":1632173261000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["Intragroup sparsity for efficient inference"],"prefix":"10.1109","author":[{"given":"Zilin","family":"Yu","sequence":"first","affiliation":[]},{"given":"Chao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Xundong","family":"Wu","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/309"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.205"},{"key":"ref33","article-title":"Dnn dataflow choice is overrated","author":"yang","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref32","article-title":"Understanding the disharmony between dropout and batch normalization by variance shift","author":"li","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref30","first-page":"1929","article-title":"Dropout: a simple way to prevent neural networks from overfitting","volume":"15","author":"srivastava","year":"2014","journal-title":"The Journal of Machine Learning Research"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.29007\/7ksb"},{"key":"ref36","article-title":"Sgdr: Stochastic gradient descent with warm restarts","author":"loshchilov","year":"2016","journal-title":"ArXiv Preprint"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"journal-title":"Learning multiple layers of features from tiny images","year":"2009","author":"krizhevsky","key":"ref34"},{"key":"ref10","article-title":"Pruning filters for efficient convnets","author":"li","year":"2016","journal-title":"ArXiv Preprint"},{"key":"ref11","first-page":"2074","article-title":"Learning structured sparsity in deep neural networks","author":"wen","year":"2016","journal-title":"Advances in neural information processing systems"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3307650.3322258"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2869974"},{"key":"ref14","first-page":"988","article-title":"Exclusive lasso for multi-task feature selection","author":"zhou","year":"0","journal-title":"Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics"},{"key":"ref15","first-page":"8057","article-title":"Improved expressivity through dendritic neural networks","author":"wu","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref16","article-title":"Cross-channel intragroup sparsity neural network","author":"yu","year":"2019","journal-title":"ArXiv Preprint"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1145\/3289602.3293898"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3007787.3001163"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/3079856.3080254"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-018-04316-3"},{"key":"ref4","article-title":"Accelerating sparsity in the nvidia ampere architecture","author":"pool","year":"2020","journal-title":"GTC 2020"},{"key":"ref27","article-title":"Targeted dropout","author":"gomez","year":"0","journal-title":"NIPS 2018 CDNNRIA Workshop"},{"key":"ref3","article-title":"Exploring the regularity of sparse structure in convolutional neural networks","author":"mao","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref29","article-title":"Parameter efficient training of deep convolutional neural networks by dynamic sparse reparameterization","author":"mostafa","year":"2019","journal-title":"ArXiv Preprint"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015676"},{"key":"ref8","first-page":"2752","article-title":"Con-densenet: An efficient densenet using learned group convolutions","author":"huang","year":"0","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref7","first-page":"5203","article-title":"Channelnets: Compact and efficient convolutional neural networks via channel-wise convolutions","author":"gao","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref2","article-title":"Mobilenets: Efficient convolutional neural networks for mobile vision applications","author":"howard","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"ref1","first-page":"1135","article-title":"Learning both weights and connections for efficient neural network","author":"han","year":"2015","journal-title":"Advances in neural information processing systems"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO.2016.7783723"},{"key":"ref22","article-title":"cudnn: Efficient primitives for deep learning","author":"chetlur","year":"2014","journal-title":"ArXiv Preprint"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO.2018.00011"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2020.2976475"},{"key":"ref23","article-title":"The indirect convolution algorithm","author":"dukhan","year":"2019","journal-title":"ArXiv Preprint"},{"key":"ref26","article-title":"To prune, or not to prune: exploring the efficacy of pruning for model compression","author":"zhu","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1145\/800076.802486"}],"event":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2021,7,18]]},"location":"Shenzhen, China","end":{"date-parts":[[2021,7,22]]}},"container-title":["2021 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9533266\/9533267\/09533623.pdf?arnumber=9533623","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:46:00Z","timestamp":1652197560000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9533623\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,18]]},"references-count":39,"URL":"https:\/\/doi.org\/10.1109\/ijcnn52387.2021.9533623","relation":{},"subject":[],"published":{"date-parts":[[2021,7,18]]}}}