{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,26]],"date-time":"2025-04-26T05:06:45Z","timestamp":1745644005659},"reference-count":11,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,4,18]],"date-time":"2022-04-18T00:00:00Z","timestamp":1650240000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,4,18]],"date-time":"2022-04-18T00:00:00Z","timestamp":1650240000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,4,18]]},"DOI":"10.1109\/vlsi-dat54769.2022.9768087","type":"proceedings-article","created":{"date-parts":[[2022,5,9]],"date-time":"2022-05-09T20:04:50Z","timestamp":1652126690000},"page":"1-4","source":"Crossref","is-referenced-by-count":3,"title":["Hardware-Friendly Progressive Pruning Framework for CNN Model Compression using Universal Pattern Sets"],"prefix":"10.1109","author":[{"given":"Wei-Cheng","family":"Chou","sequence":"first","affiliation":[{"name":"Institute of Electronics, National Yang Ming Chiao Tung University,Hsinchu,Taiwan"}]},{"given":"Cheng-Wei","family":"Huang","sequence":"additional","affiliation":[{"name":"Institute of Electronics, National Yang Ming Chiao Tung University,Hsinchu,Taiwan"}]},{"given":"Juinn-Dar","family":"Huang","sequence":"additional","affiliation":[{"name":"Institute of Electronics, National Yang Ming Chiao Tung University,Hsinchu,Taiwan"}]}],"member":"263","reference":[{"key":"ref4","first-page":"2074","article-title":"Learning structured sparsity in deep neural networks","author":"wen","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst (NIPS)"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3297858.3304076"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/DAC18072.2020.9218498"},{"journal-title":"Ss-auto A singleshot automatic structured weight pruning framework of dnns with ultra-high efficiency","year":"2020","author":"li","key":"ref6"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5954"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.155"},{"journal-title":"Blk-rew A unifed block-based DNN pruning framework using reweighted regularization method","year":"2020","author":"ma","key":"ref8"},{"journal-title":"Structadmm A systematic high-efficiency framework of structured weight pruning for dnns","year":"2018","author":"zhang","key":"ref7"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_12"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/3373376.3378534"},{"key":"ref1","first-page":"1135","article-title":"Learning both weights and connections for efficient neural network","author":"han","year":"2015","journal-title":"Proc Adv Neural Inf Process Syst (NIPS)"}],"event":{"name":"2022 International Symposium on VLSI Design, Automation and Test (VLSI-DAT)","start":{"date-parts":[[2022,4,18]]},"location":"Hsinchu, Taiwan","end":{"date-parts":[[2022,4,21]]}},"container-title":["2022 International Symposium on VLSI Design, Automation and Test (VLSI-DAT)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9768045\/9768046\/09768087.pdf?arnumber=9768087","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T21:11:56Z","timestamp":1655759516000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9768087\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,18]]},"references-count":11,"URL":"https:\/\/doi.org\/10.1109\/vlsi-dat54769.2022.9768087","relation":{},"subject":[],"published":{"date-parts":[[2022,4,18]]}}}