{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T12:16:18Z","timestamp":1767183378586,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,7,10]],"date-time":"2022-07-10T00:00:00Z","timestamp":1657411200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"SRC Global Research Collaboration (GRC)"},{"name":"NSF","award":["1911095, 1826967, 2100237, 2112167"],"award-info":[{"award-number":["1911095, 1826967, 2100237, 2112167"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,7,10]]},"DOI":"10.1145\/3489517.3530422","type":"proceedings-article","created":{"date-parts":[[2022,8,23]],"date-time":"2022-08-23T23:19:29Z","timestamp":1661296769000},"page":"223-228","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["PatterNet"],"prefix":"10.1145","author":[{"given":"Behnam","family":"Khaleghi","sequence":"first","affiliation":[{"name":"UC San Diego"}]},{"given":"Uday","family":"Mallappa","sequence":"additional","affiliation":[{"name":"UC San Diego"}]},{"given":"Duygu","family":"Yaldiz","sequence":"additional","affiliation":[{"name":"UC San Diego"}]},{"given":"Haichao","family":"Yang","sequence":"additional","affiliation":[{"name":"UC San Diego"}]},{"given":"Monil","family":"Shah","sequence":"additional","affiliation":[{"name":"UC San Diego"}]},{"given":"Jaeyoung","family":"Kang","sequence":"additional","affiliation":[{"name":"UC San Diego"}]},{"given":"Tajana","family":"Rosing","sequence":"additional","affiliation":[{"name":"UC San Diego"}]}],"member":"320","published-online":{"date-parts":[[2022,8,23]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"An analysis of deep neural network models for practical applications,\" arXiv preprint arXiv:1605.07678","author":"Canziani A.","year":"2016","unstructured":"A. Canziani, A. Paszke, and E. Culurciello, \"An analysis of deep neural network models for practical applications,\" arXiv preprint arXiv:1605.07678, 2016."},{"key":"e_1_3_2_1_2_1","first-page":"525","volume-title":"Xnor-net: Imagenet classification using binary convolutional neural networks,\" in European conference on computer vision","author":"Rastegari M.","year":"2016","unstructured":"M. Rastegari et al., \"Xnor-net: Imagenet classification using binary convolutional neural networks,\" in European conference on computer vision, pp. 525--542, Springer, 2016."},{"key":"e_1_3_2_1_3_1","volume-title":"Trained ternary quantization,\" arXiv preprint arXiv:1612.01064","author":"Zhu C.","year":"2016","unstructured":"C. Zhu, S. Han, H. Mao, and W. J. Dally, \"Trained ternary quantization,\" arXiv preprint arXiv:1612.01064, 2016."},{"key":"e_1_3_2_1_4_1","volume-title":"Incremental network quantization: Towards lossless cnns with low-precision weights,\" arXiv preprint arXiv:1702.03044","author":"Zhou A.","year":"2017","unstructured":"A. Zhou, A. Yao, Y. Guo, L. Xu, and Y. Chen, \"Incremental network quantization: Towards lossless cnns with low-precision weights,\" arXiv preprint arXiv:1702.03044, 2017."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"A. Ren T. Zhang et al. \"Admm-nn: An algorithm-hardware co-design framework of dnns using alternating direction methods of multipliers \" in International Conference on Architectural Support for Programming Languages and Operating Systems pp. 925--938 2019.","DOI":"10.1145\/3297858.3304076"},{"key":"e_1_3_2_1_6_1","volume-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding,\" arXiv preprint arXiv:1510.00149","author":"Han S.","year":"2015","unstructured":"S. Han, H. Mao, and W. J. Dally, \"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding,\" arXiv preprint arXiv:1510.00149, 2015."},{"key":"e_1_3_2_1_7_1","volume-title":"A unified framework of dnn weight pruning and weight clustering\/quantization using admm,\" arXiv preprint arXiv:1811.01907","author":"Ye S.","year":"2018","unstructured":"S. Ye, T. Zhang, et al., \"A unified framework of dnn weight pruning and weight clustering\/quantization using admm,\" arXiv preprint arXiv:1811.01907, 2018."},{"key":"e_1_3_2_1_8_1","volume-title":"Pruning filters for efficient convnets,\" arXiv preprint arXiv:1608.08710","author":"Li H.","year":"2016","unstructured":"H. Li, A. Kadav, I. Durdanovic, H. Samet, and H. P. Graf, \"Pruning filters for efficient convnets,\" arXiv preprint arXiv:1608.08710, 2016."},{"key":"e_1_3_2_1_9_1","first-page":"1529","article-title":"Hrank: Filter pruning using high-rank feature map","author":"Lin M.","year":"2020","unstructured":"M. Lin et al., \"Hrank: Filter pruning using high-rank feature map,\" in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1529--1538, 2020.","journal-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition"},{"key":"e_1_3_2_1_10_1","first-page":"208","volume-title":"Mix and match: A novel fpga-centric deep neural network quantization framework\" in International Symposium on High-Performance Computer Architecture (HPCA)","author":"Chang S.-E.","year":"2021","unstructured":"S.-E. Chang, Y. Li, et al., \"Mix and match: A novel fpga-centric deep neural network quantization framework\" in International Symposium on High-Performance Computer Architecture (HPCA), pp. 208--220, IEEE, 2021."},{"key":"e_1_3_2_1_11_1","first-page":"1389","article-title":"Channel pruning for accelerating very deep neural networks","author":"He Y.","year":"2017","unstructured":"Y. He, X. Zhang, and J. Sun, \"Channel pruning for accelerating very deep neural networks,\" in Proceedings of the IEEE international conference on computer vision, pp. 1389--1397, 2017.","journal-title":"Proceedings of the IEEE international conference on computer vision"},{"key":"e_1_3_2_1_12_1","first-page":"674","volume-title":"Ucnn: Exploiting computational reuse in deep neural networks via weight repetition,\" in International Symposium on Computer Architecture (ISCA)","author":"Hegde K.","year":"2018","unstructured":"K. Hegde et al., \"Ucnn: Exploiting computational reuse in deep neural networks via weight repetition,\" in International Symposium on Computer Architecture (ISCA), pp. 674--687, 2018."},{"key":"e_1_3_2_1_13_1","first-page":"894","volume-title":"Fusekna: Fused kernel convolution based accelerator for deep neural networks,\" in International Symposium on High-Performance Computer Architecture","author":"Yang J.","year":"2021","unstructured":"J. Yang et al., \"Fusekna: Fused kernel convolution based accelerator for deep neural networks,\" in International Symposium on High-Performance Computer Architecture, pp. 894--907, 2021."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1002\/9780470316801.ch2"},{"key":"e_1_3_2_1_15_1","first-page":"1543","volume-title":"Automation & Test in Europe Conference & Exhibition (DATE)","author":"Poremba M.","year":"2015","unstructured":"M. Poremba, S. Mittal, D. Li, J. S. Vetter, and Y. Xie, \"Destiny: A tool for modeling emerging 3d nvm and edram caches,\" in 2015 Design, Automation & Test in Europe Conference & Exhibition (DATE), pp. 1543--1546, IEEE, 2015."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3140659.3080254"}],"event":{"name":"DAC '22: 59th ACM\/IEEE Design Automation Conference","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE CEDA"],"location":"San Francisco California","acronym":"DAC '22"},"container-title":["Proceedings of the 59th ACM\/IEEE Design Automation Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3489517.3530422","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3489517.3530422","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3489517.3530422","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:39Z","timestamp":1750191519000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3489517.3530422"}},"subtitle":["explore and exploit filter patterns for efficient deep neural networks"],"short-title":[],"issued":{"date-parts":[[2022,7,10]]},"references-count":16,"alternative-id":["10.1145\/3489517.3530422","10.1145\/3489517"],"URL":"https:\/\/doi.org\/10.1145\/3489517.3530422","relation":{},"subject":[],"published":{"date-parts":[[2022,7,10]]},"assertion":[{"value":"2022-08-23","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}