{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,23]],"date-time":"2025-10-23T00:06:17Z","timestamp":1761177977003,"version":"build-2065373602"},"publisher-location":"ISCA","reference-count":0,"publisher":"ISCA","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"DOI":"10.21437\/interspeech.2025-503","type":"proceedings-article","created":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T11:47:41Z","timestamp":1761133661000},"page":"1988-1992","source":"Crossref","is-referenced-by-count":0,"title":["Ultra-Low Bit Post-Training Quantization of Large Speech Models via K-Means Clustering and Mixed Precision Allocation"],"prefix":"10.21437","author":[{"given":"Tianteng","family":"Gu","sequence":"first","affiliation":[]},{"given":"Bei","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Haoyu","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yanmin","family":"Qian","sequence":"additional","affiliation":[]}],"member":"8866","published-online":{"date-parts":[[2025,8,17]]},"event":{"name":"Interspeech 2025","acronym":"interspeech_2025"},"container-title":["Interspeech 2025"],"original-title":[],"deposited":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T11:48:16Z","timestamp":1761133696000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.isca-archive.org\/interspeech_2025\/gu25b_interspeech.html"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,17]]},"references-count":0,"URL":"https:\/\/doi.org\/10.21437\/interspeech.2025-503","relation":{},"subject":[],"published":{"date-parts":[[2025,8,17]]}}}