{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T14:37:58Z","timestamp":1769179078626,"version":"3.49.0"},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,5]]},"DOI":"10.1109\/spawc48557.2020.9154309","type":"proceedings-article","created":{"date-parts":[[2020,8,3]],"date-time":"2020-08-03T22:12:25Z","timestamp":1596492745000},"page":"1-5","source":"Crossref","is-referenced-by-count":18,"title":["Analog Compression and Communication for Federated Learning over Wireless MAC"],"prefix":"10.1109","author":[{"given":"Afshin","family":"Abdi","sequence":"first","affiliation":[]},{"given":"Yashas Malur","family":"Saidutta","sequence":"additional","affiliation":[]},{"given":"Faramarz","family":"Fekri","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref32","article-title":"Error Compensated Quantized SGD and its Applications to Large-scale Distributed Optimization","author":"wu","year":"2018","journal-title":"ICML"},{"key":"ref31","article-title":"Error Feedback Fixes SignSGD and other Gradient Compression Schemes","author":"karimireddy","year":"2019","journal-title":"arXiv preprint arXiv 1901 04217"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT.2019.8849334"},{"key":"ref10","doi-asserted-by":"crossref","first-page":"2219","DOI":"10.1109\/ICASSP.2015.7178365","article-title":"Reducing communication overhead in distributed learning by an order of magnitude (almost)","author":"\u00f8land","year":"2015","journal-title":"2015 IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP)"},{"key":"ref11","first-page":"1707","article-title":"QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding","author":"alistarh","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref12","first-page":"560","article-title":"SignSGD: Compressed optimisation for non-convex problems","author":"bernstein","year":"2018","journal-title":"Proceedings of the 35th International Conference on Machine Learning ser Proceedings of Machine Learning Research"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/MLHPC.2016.004"},{"key":"ref14","article-title":"Fast convergence rates of distributed subgradient methods with adaptive quantization","author":"doan","year":"2018","journal-title":"arXiv preprint arXiv 1810 10053"},{"key":"ref15","article-title":"Nested dithered quantization for communication reduction in distributed training","author":"abdi","year":"2019","journal-title":"arXiv preprint arXiv 1904 01870"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/SPAWC.2019.8815453"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5706"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5707"},{"key":"ref19","first-page":"10","article-title":"Scalable distributed DNN training using commodity GPU cloud computing","volume":"7","author":"strom","year":"2015","journal-title":"InterSpeech"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2019.8737464"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/3298981"},{"key":"ref27","article-title":"Over-the-air computation systems: Optimization, analysis and scaling laws","author":"liu","year":"2019","journal-title":"arXiv preprint arXiv 1909 01771"},{"key":"ref3","article-title":"Federated optimization: Distributed machine learning for on-device intelligence","author":"kone?n?","year":"2016","journal-title":"arXiv preprint arXiv 1610 09756"},{"key":"ref6","first-page":"1","article-title":"Local SGD Converges Fast and Communicates Little","author":"stich","year":"2019","journal-title":"ICLRE"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/SPAWC.2019.8815402"},{"key":"ref5","article-title":"Federated learning with non-iid data","author":"zhao","year":"2018","journal-title":"arXiv preprint arXiv 1806 00582"},{"key":"ref8","article-title":"Robust and communication-efficient federated learning from non-iid data","author":"sattler","year":"2019","journal-title":"arXiv preprint arXiv 1903 00066"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015693"},{"key":"ref2","article-title":"Federated learning: Strategies for improving communication efficiency","author":"konecn\u00fd","year":"2016","journal-title":"Proc NIPS Workshop Private Multi-Party Mach Learn"},{"key":"ref9","first-page":"1058","article-title":"1-bit stochastic gradient descent and its application to data-parallel distributed training of speech DNNs","author":"seide","year":"2014","journal-title":"InterSpeech"},{"key":"ref1","article-title":"Federated learning of deep networks using model averaging","author":"mcmahan","year":"2016","journal-title":"arXiv preprint arXiv 1602 05629v1"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1045"},{"key":"ref22","first-page":"4452","article-title":"Sparsified sgd with memory","author":"stich","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref21","article-title":"SparCML: High-performance sparse communication for machine learning","author":"renggli","year":"2018","journal-title":"arXiv preprint arXiv 1802 10363"},{"key":"ref24","first-page":"5977","article-title":"The convergence of sparsified gradient methods","author":"alistarh","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref23","first-page":"1306","article-title":"Gradient sparsification for communication-efficient distributed optimization","author":"wangni","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref26","article-title":"On analog gradient descent learning over multiple access fading channels","author":"sery","year":"2019","journal-title":"arXiv preprint arXiv 1908 07463"},{"key":"ref25","article-title":"Broadband analog aggregation for low-latency federated edge learning","author":"zhu","year":"2018","journal-title":"arXiv preprint arXiv 1812 11494 v3"}],"event":{"name":"2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications (SPAWC)","location":"Atlanta, GA, USA","start":{"date-parts":[[2020,5,26]]},"end":{"date-parts":[[2020,5,29]]}},"container-title":["2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications (SPAWC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9144330\/9153881\/09154309.pdf?arnumber=9154309","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,30]],"date-time":"2022-06-30T15:18:19Z","timestamp":1656602299000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9154309\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,5]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/spawc48557.2020.9154309","relation":{},"subject":[],"published":{"date-parts":[[2020,5]]}}}