{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T14:57:16Z","timestamp":1773154636426,"version":"3.50.1"},"publisher-location":"ISCA","reference-count":0,"publisher":"ISCA","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"DOI":"10.21437\/interspeech.2021-248","type":"proceedings-article","created":{"date-parts":[[2021,8,27]],"date-time":"2021-08-27T05:59:39Z","timestamp":1630043979000},"page":"4568-4572","source":"Crossref","is-referenced-by-count":58,"title":["PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation"],"prefix":"10.21437","author":[{"given":"Jangho","family":"Kim","sequence":"first","affiliation":[]},{"given":"Simyung","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Nojun","family":"Kwak","sequence":"additional","affiliation":[]}],"member":"8866","published-online":{"date-parts":[[2021,8,30]]},"event":{"name":"Interspeech 2021","acronym":"interspeech_2021"},"container-title":["Interspeech 2021"],"original-title":[],"deposited":{"date-parts":[[2024,1,16]],"date-time":"2024-01-16T10:19:58Z","timestamp":1705400398000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.isca-archive.org\/interspeech_2021\/kim21m_interspeech.html"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,30]]},"references-count":0,"URL":"https:\/\/doi.org\/10.21437\/interspeech.2021-248","relation":{},"subject":[],"published":{"date-parts":[[2021,8,30]]}}}