{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T04:37:40Z","timestamp":1773376660471,"version":"3.50.1"},"reference-count":29,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,6,25]],"date-time":"2023-06-25T00:00:00Z","timestamp":1687651200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,6,25]],"date-time":"2023-06-25T00:00:00Z","timestamp":1687651200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,6,25]]},"DOI":"10.1109\/isit54713.2023.10206664","type":"proceedings-article","created":{"date-parts":[[2023,8,22]],"date-time":"2023-08-22T13:31:15Z","timestamp":1692711075000},"page":"719-724","source":"Crossref","is-referenced-by-count":2,"title":["Optimal Compression of Unit Norm Vectors in the High Distortion Regime"],"prefix":"10.1109","author":[{"given":"Heng","family":"Zhu","sequence":"first","affiliation":[{"name":"University of California,San Diego"}]},{"given":"Avishek","family":"Ghosh","sequence":"additional","affiliation":[{"name":"Indian Institute of Technology, Bombay"}]},{"given":"Arya","family":"Mazumdar","sequence":"additional","affiliation":[{"name":"University of California,San Diego"}]}],"member":"263","reference":[{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1002\/0471219282.eot142"},{"key":"ref12","first-page":"2197","article-title":"vqsgd: Vector quantized stochastic gradient descent","author":"gandikota","year":"2021","journal-title":"International Conference on Artificial Intelligence and Statistics"},{"key":"ref15","first-page":"3312","article-title":"Breaking the communication-privacy-accuracy trilemma","volume":"33","author":"chen","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref14","author":"cohen","year":"1997","journal-title":"Covering Codes"},{"key":"ref11","first-page":"1299","article-title":"Gradient sparsification for communication-efficient distributed optimization","author":"wangni","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/JSAIT.2021.3105076"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-84882-903-9"},{"key":"ref1","volume":"159","author":"gersho","year":"2012","journal-title":"Vector Quantization and Signal Compression"},{"key":"ref17","article-title":"Local SGD converges fast and communicates little","author":"stich","year":"2019","journal-title":"International Conference on Learning Representations"},{"key":"ref16","first-page":"3056","article-title":"The fundamental price of secure aggregation in differentially private federated learning","author":"chen","year":"2022","journal-title":"International Conference on Machine Learning"},{"key":"ref19","article-title":"Communication efficient distributed optimization using an approximate newton-type method","author":"shamir","year":"2013","journal-title":"CoRR"},{"key":"ref18","first-page":"5132","article-title":"Scaffold: Stochastic controlled averaging for federated learning","author":"karimireddy","year":"2020","journal-title":"International Conference on Machine Learning"},{"key":"ref24","first-page":"1509","article-title":"Terngrad: Ternary gradients to reduce communication in distributed deep learning","author":"wen","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref23","article-title":"Introduction to the non-asymptotic analysis of random matrices","author":"vershynin","year":"2010"},{"key":"ref26","article-title":"Randomized distributed mean estimation: Accuracy vs. communication","volume":"4","author":"kone?n?","year":"2018","journal-title":"Front Appl Math Stat"},{"key":"ref25","first-page":"4035","article-title":"Zipml: Training linear models with end-to-end low precision, and a little bit of deep learning","author":"zhang","year":"2017","journal-title":"International Conference on Machine Learning"},{"key":"ref20","article-title":"Giant: Globally improved approximate newton method for distributed optimization","author":"wang","year":"2017"},{"key":"ref22","article-title":"Escaping saddle points in distributed newton&#x2019;s method with communication efficiency and byzantine resilience","author":"ghosh","year":"2021","journal-title":"CoRR"},{"key":"ref21","article-title":"Distributed newton can communicate less and resist byzantine workers","author":"ghosh","year":"2020","journal-title":"Proceedings of the 34th International Conference on Neural Information Processing Systems"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT44484.2020.9174075"},{"key":"ref27","first-page":"1399","article-title":"Ratq: A universal fixed-length quantizer for stochastic optimization","author":"mayekar","year":"2020","journal-title":"International Conference on Artificial Intelligence and Statistics"},{"key":"ref29","first-page":"3329","article-title":"Distributed mean estimation with limited communication","author":"suresh","year":"2017","journal-title":"International Conference on Machine Learning"},{"key":"ref8","first-page":"4447","article-title":"Sparsified SGD with memory","author":"stich","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref7","first-page":"1709","article-title":"QSGD: Communication-efficient SGD via gradient quantization and encoding","author":"alistarh","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref9","first-page":"3252","article-title":"Error feedback fixes SignSGD and other gradient compression schemes","author":"karimireddy","year":"2019","journal-title":"International Conference on Machine Learning"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1561\/2200000083"},{"key":"ref3","author":"sayood","year":"2017","journal-title":"Introduction to Data Compression"},{"key":"ref6","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","author":"mcmahan","year":"2017","journal-title":"Artificial Intelligence and Statistics"},{"key":"ref5","article-title":"Federated learning: Strategies for improving communication efficiency","author":"kone?n?","year":"2016"}],"event":{"name":"2023 IEEE International Symposium on Information Theory (ISIT)","location":"Taipei, Taiwan","start":{"date-parts":[[2023,6,25]]},"end":{"date-parts":[[2023,6,30]]}},"container-title":["2023 IEEE International Symposium on Information Theory (ISIT)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10206429\/10206441\/10206664.pdf?arnumber=10206664","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T20:27:25Z","timestamp":1773347245000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10206664\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,25]]},"references-count":29,"URL":"https:\/\/doi.org\/10.1109\/isit54713.2023.10206664","relation":{},"subject":[],"published":{"date-parts":[[2023,6,25]]}}}