{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:27:49Z","timestamp":1777656469157,"version":"3.51.4"},"reference-count":41,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"am","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61873118"],"award-info":[{"award-number":["61873118"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shenzhen Committee on Science and Innovations","award":["GJHZ20180411143603361"],"award-info":[{"award-number":["GJHZ20180411143603361"]}]},{"DOI":"10.13039\/501100007162","name":"Guangdong Science and Technology Department","doi-asserted-by":"publisher","award":["2018A050506003"],"award-info":[{"award-number":["2018A050506003"]}],"id":[{"id":"10.13039\/501100007162","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004543","name":"China Scholarship Council","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100004543","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key-Area Research and Development Program of Guangdong Province","award":["2018B010107002"],"award-info":[{"award-number":["2018B010107002"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61751205"],"award-info":[{"award-number":["61751205"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["1901134"],"award-info":[{"award-number":["1901134"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2022,4,1]]},"DOI":"10.1109\/tpami.2020.3033286","type":"journal-article","created":{"date-parts":[[2020,10,23]],"date-time":"2020-10-23T19:49:20Z","timestamp":1603482560000},"page":"2031-2044","source":"Crossref","is-referenced-by-count":90,"title":["Lazily Aggregated Quantized Gradient Innovation for Communication-Efficient Federated Learning"],"prefix":"10.1109","volume":"44","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2707-8751","authenticated-orcid":false,"given":"Jun","family":"Sun","sequence":"first","affiliation":[{"name":"State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3477-1439","authenticated-orcid":false,"given":"Tianyi","family":"Chen","sequence":"additional","affiliation":[{"name":"Department of Electrical, Computer, and Systems Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0196-0260","authenticated-orcid":false,"given":"Georgios B.","family":"Giannakis","sequence":"additional","affiliation":[{"name":"Department of Electrical and Computer Engineering and the Digital Technology Center, University of Minnesota, Minneapolis, MN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1602-8986","authenticated-orcid":false,"given":"Qinmin","family":"Yang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8288-3833","authenticated-orcid":false,"given":"Zaiyue","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Mechanical and Energy Engineering, Southern University of Science and Technology, Shenzhen, China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"3370","article-title":"Communication-efficient distributed learning via lazily aggregated quantized gradients","volume-title":"Proc. 33rd Int. Conf. Neural Inf. Process. Syst.","author":"Sun"},{"key":"ref2","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","volume-title":"Proc. 20th Int. Conf. Artif. Intell. Statist.","author":"McMahan"},{"key":"ref3","article-title":"Federated learning: Strategies for improving communication efficiency","author":"Kone\u010dn\u1ef3","year":"2016"},{"key":"ref4","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","volume-title":"Proc. 20th Int. Conf. Artif. Intell. Statist.","author":"McMahan"},{"key":"ref5","first-page":"19","article-title":"Communication efficient distributed machine learning with the parameter server","volume-title":"Proc. 27th Int. Conf. Neural Inf. Process. Syst.","author":"Li"},{"key":"ref6","first-page":"3252","article-title":"Error feedback fixes signSGD and other gradient compression schemes","volume-title":"Proc. 36th Int. Conf. Mach. Learn.","author":"Karimireddy"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2011.2171686"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2014-274"},{"key":"ref9","first-page":"559","article-title":"SignSGD: Compressed optimisation for non-convex problems","volume-title":"Proc. 35th Int. Conf. Mach. Learn.","author":"Bernstein"},{"key":"ref10","first-page":"1709","article-title":"QSGD: Communication-efficient SGD via gradient quantization and encoding","volume-title":"Proc. 31st Int. Conf. Neural Inf. Process. Syst.","author":"Alistarh"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/IEEECONF44664.2019.9049052"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.5555\/3294771.3294915"},{"key":"ref13","first-page":"4035","article-title":"ZipML: Training linear models with end-to-end low precision, and a little bit of deep learning","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Zhang"},{"key":"ref14","article-title":"Error compensated quantized SGD and its applications to large-scale distributed optimization","author":"Wu","year":"2018"},{"key":"ref15","article-title":"Distributed learning with compressed gradient differences","author":"Mishchenko","year":"2019"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2015-354"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1045"},{"key":"ref18","first-page":"1","article-title":"Deep gradient compression: Reducing the communication bandwidth for distributed training","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Lin"},{"key":"ref19","first-page":"4447","article-title":"Sparsified SGD with memory","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Stich"},{"key":"ref20","first-page":"5973","article-title":"The convergence of sparsified gradient methods","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Alistarh"},{"key":"ref21","first-page":"1299","article-title":"Gradient sparsification for communication-efficient distributed optimization","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Wangni"},{"key":"ref22","first-page":"9850","article-title":"ATOMO: Communication-efficient learning via atomic sparsification","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref23","first-page":"2525","article-title":"A linear speedup analysis of distributed deep learning with sparse and quantized communication","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Jiang"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.3389\/fams.2018.00062"},{"key":"ref25","volume-title":"Computer Networks: A Systems Approach","author":"Peterson","year":"2007"},{"key":"ref26","first-page":"1000","article-title":"Communication-efficient distributed optimization using an approximate newton-type method","volume-title":"Proc. 31st Int. Conf. Mach. Learn.","author":"Shamir"},{"key":"ref27","first-page":"362","article-title":"DiSCO: Distributed optimization for self-concordant empirical loss","volume-title":"Proc. 32nd Int. Conf. Mach. Learn.","author":"Zhang"},{"key":"ref28","first-page":"685","article-title":"Deep learning with elastic averaging SGD","volume-title":"Proc. 28th Int. Conf. Neural Inf. Process. Syst.","author":"Zhang"},{"key":"ref29","article-title":"On the computation and communication complexity of parallel SGD with dynamic batch sizes for stochastic non-convex optimization","author":"Yu","year":"2019"},{"key":"ref30","first-page":"5050","article-title":"LAG: Lazily aggregated gradient for communication-efficient distributed learning","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Chen"},{"key":"ref31","article-title":"Communication-efficient distributed reinforcement learning","author":"Chen","year":"2018"},{"key":"ref32","article-title":"Cooperative SGD: A unified framework for the design and analysis of communication-efficient SGD algorithms","author":"Wang","year":"2018"},{"key":"ref33","first-page":"1756","article-title":"Communication complexity of distributed convex learning and optimization","volume-title":"Proc. 28th Int. Conf. Neural Inf. Process. Syst.","author":"Arjevani"},{"key":"ref34","first-page":"4120","article-title":"Asynchronous stochastic gradient descent with delay compensation","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Zheng"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1080\/01621459.2018.1429274"},{"key":"ref36","first-page":"14 774","article-title":"Deep leakage from gradients","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Zhu"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/431"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1137\/15M1049695"},{"key":"ref39","article-title":"MNIST handwritten digit database","volume-title":"AT&T Labs","author":"LeCun","year":"2010"},{"key":"ref40","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2014"},{"key":"ref41","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/ieeexplore.ieee.org\/ielam\/34\/9729045\/9238427-aam.pdf","content-type":"application\/pdf","content-version":"am","intended-application":"syndication"},{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/9729045\/09238427.pdf?arnumber=9238427","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,9]],"date-time":"2024-01-09T23:13:43Z","timestamp":1704842023000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9238427\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,1]]},"references-count":41,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2020.3033286","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,1]]}}}