{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T22:16:09Z","timestamp":1772576169136,"version":"3.50.1"},"reference-count":46,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"am","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1806154"],"award-info":[{"award-number":["CCF-1806154"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1901199"],"award-info":[{"award-number":["CCF-1901199"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-2007911"],"award-info":[{"award-number":["CCF-2007911"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000006","name":"Office of Naval Research","doi-asserted-by":"publisher","award":["N00014-18-1-2142"],"award-info":[{"award-number":["N00014-18-1-2142"]}],"id":[{"id":"10.13039\/100000006","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000006","name":"Office of Naval Research","doi-asserted-by":"publisher","award":["N00014-19-1-2404"],"award-info":[{"award-number":["N00014-19-1-2404"]}],"id":[{"id":"10.13039\/100000006","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000183","name":"Army Research Office","doi-asserted-by":"publisher","award":["W911NF-18-1-0303"],"award-info":[{"award-number":["W911NF-18-1-0303"]}],"id":[{"id":"10.13039\/100000183","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Signal Process."],"published-print":{"date-parts":[[2020]]},"DOI":"10.1109\/tsp.2020.3005291","type":"journal-article","created":{"date-parts":[[2020,6,26]],"date-time":"2020-06-26T22:27:55Z","timestamp":1593210475000},"page":"3976-3989","source":"Crossref","is-referenced-by-count":14,"title":["Convergence of Distributed Stochastic Variance Reduced Methods Without Sampling Extra Data"],"prefix":"10.1109","volume":"68","author":[{"given":"Shicong","family":"Cen","sequence":"first","affiliation":[]},{"given":"Huishuai","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6766-5459","authenticated-orcid":false,"given":"Yuejie","family":"Chi","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Tie-Yan","family":"Liu","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2018.2872003"},{"key":"ref38","first-page":"4855","article-title":"${D}^{2}$: Decentralized training over decentralized data","author":"tang","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref33","first-page":"1509","article-title":"Terngrad: Ternary gradients to reduce communication in distributed deep learning","author":"wen","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref32","first-page":"1058","author":"seide","year":"0","journal-title":"Proc Annu Conf Int Speech Commun Assoc"},{"key":"ref31","first-page":"559","article-title":"SignSDG: Compressed optimisation for non-convex problems","author":"bernstein","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref30","first-page":"1709","article-title":"QSGD: Communication-efficient SGD via gradient quantization and encoding","author":"alistarh","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref37","first-page":"1299","article-title":"Gradient sparsification for communication-efficient distributed optimization","author":"wangni","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref36","author":"lin","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref35","first-page":"5330","article-title":"Can decentralized algorithms outperform centralized algorithms? A case study for decentralized parallel stochastic gradient descent","author":"lian","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref34","first-page":"5973","article-title":"The convergence of sparsified gradient methods","author":"alistarh","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref10","article-title":"Finite-sum smooth optimization with SARAH","author":"nguyen","year":"2019"},{"key":"ref40","first-page":"2165","article-title":"DSA: Decentralized double stochastic averaging gradient algorithm","volume":"17","author":"mokhtari","year":"2016","journal-title":"J Mach Learn Res"},{"key":"ref11","first-page":"5975","article-title":"A simple stochastic variance reduced algorithm with fast convergence rates","author":"zhou","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref12","first-page":"2928","article-title":"SCOPE: Scalable composite optimization for learning on spark","author":"zhao","year":"0","journal-title":"Proc 31st AAAI Conf Artif Intell"},{"key":"ref13","first-page":"6552","article-title":"Proximal SCOPE for distributed sparse learning","author":"zhao","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1561\/2200000016"},{"key":"ref15","volume":"23","author":"bertsekas","year":"1989","journal-title":"Parallel and Distributed Computation Numerical Methods"},{"key":"ref16","first-page":"1000","article-title":"Communication-efficient distributed optimization using an approximate Newton-type method","author":"shamir","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref17","first-page":"362","article-title":"DiSCO: Distributed optimization for self-concordant empirical loss","author":"zhang","year":"2015","journal-title":"Int Conf Mach Learn"},{"key":"ref18","first-page":"2338","article-title":"Giant: Globally improved approximate newton method for distributed optimization","author":"wang","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref19","first-page":"8590","article-title":"CoCoA: A general framework for communication-efficient distributed optimization","volume":"18","author":"smith","year":"2017","journal-title":"J Mach Learn Res"},{"key":"ref28","first-page":"687","article-title":"Spider: Near-optimal non-convex optimization via stochastic path-integrated differential estimator","author":"fang","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref4","first-page":"111","article-title":"Efficient distributed sgd with variance reduction","author":"goldstein","year":"0","journal-title":"Proc IEEE 16th Int Conf Data Mining"},{"key":"ref27","first-page":"3384","article-title":"A universal catalyst for first-order optimization","author":"lin","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref3","first-page":"315","article-title":"Accelerating stochastic gradient descent using predictive variance reduction","author":"johnson","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref6","first-page":"4404","article-title":"Distributed stochastic variance reduced gradient methods by sampling extra data with replacement","volume":"18","author":"lee","year":"2017","journal-title":"J Mach Learn Res"},{"key":"ref29","first-page":"2406","article-title":"SpiderBoost and momentum: Faster variance reduction algorithms","author":"wang","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref5","article-title":"Aide: Fast and communication efficient distributed optimization","author":"reddi","year":"2016"},{"key":"ref8","first-page":"1882","article-title":"Memory and communication efficient distributed stochastic optimization with minibatch prox","author":"wang","year":"0","journal-title":"Proc Conf Learn Theory"},{"key":"ref7","first-page":"46","article-title":"Without-replacement sampling for stochastic gradient methods","author":"shamir","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref2","first-page":"693","article-title":"Hogwild: A lock-free approach to parallelizing stochastic gradient descent","author":"recht","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref9","first-page":"2613","article-title":"Sarah: A novel method for machine learning problems using stochastic recursive gradient","author":"nguyen","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref1","article-title":"Federated optimization: Distributed optimization beyond the datacenter","author":"kone?n?","year":"2015"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1137\/140961791"},{"key":"ref20","article-title":"Communication-efficient accurate statistical estimation","author":"fan","year":"2019","journal-title":"arXiv 1906 04870"},{"key":"ref45","first-page":"361","article-title":"RCV1: A new benchmark collection for text categorization research","volume":"5","author":"lewis","year":"2004","journal-title":"J Mach Learn Res"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2012.6426691"},{"key":"ref21","first-page":"2595","article-title":"Parallelized stochastic gradient descent","author":"zinkevich","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1214\/17-AOS1637"},{"key":"ref24","first-page":"1646","article-title":"Saga: A fast incremental gradient method with support for non-strongly convex composite objectives","author":"defazio","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref41","first-page":"1662","author":"li","year":"0","journal-title":"Proc Int Conf Artif Intell Statist"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/s10107-016-1030-6"},{"key":"ref44","first-page":"545","article-title":"Result analysis of the nips 2003 feature selection challenge","author":"guyon","year":"0","journal-title":"Proc Advances Neural Inf Process Syst"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/3055399.3055448"},{"key":"ref43","first-page":"2038","author":"hu","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref25","first-page":"567","article-title":"Stochastic dual coordinate ascent methods for regularized loss minimization","volume":"14","author":"shalev-shwartz","year":"2013","journal-title":"J Mach Learn Res"}],"container-title":["IEEE Transactions on Signal Processing"],"original-title":[],"link":[{"URL":"https:\/\/ieeexplore.ieee.org\/ielam\/78\/8933520\/9127115-aam.pdf","content-type":"application\/pdf","content-version":"am","intended-application":"syndication"},{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/78\/8933520\/09127115.pdf?arnumber=9127115","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T14:40:38Z","timestamp":1651070438000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9127115\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"references-count":46,"URL":"https:\/\/doi.org\/10.1109\/tsp.2020.3005291","relation":{},"ISSN":["1053-587X","1941-0476"],"issn-type":[{"value":"1053-587X","type":"print"},{"value":"1941-0476","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]}}}