{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T04:39:16Z","timestamp":1773376756691,"version":"3.50.1"},"reference-count":26,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,6,25]],"date-time":"2023-06-25T00:00:00Z","timestamp":1687651200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,6,25]],"date-time":"2023-06-25T00:00:00Z","timestamp":1687651200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001321","name":"National Research Foundation","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001321","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,6,25]]},"DOI":"10.1109\/isit54713.2023.10206480","type":"proceedings-article","created":{"date-parts":[[2023,8,22]],"date-time":"2023-08-22T13:31:15Z","timestamp":1692711075000},"page":"2266-2271","source":"Crossref","is-referenced-by-count":3,"title":["S\n                    <sup>3<\/sup>\n                    GD-MV: Sparse-SignSGD with Majority Vote for Communication-Efficient Distributed Learning"],"prefix":"10.1109","author":[{"given":"Chanho","family":"Park","sequence":"first","affiliation":[{"name":"POSTECH, Electrical Engineering,Pohang,South Korea,37673"}]},{"given":"Namyoon","family":"Lee","sequence":"additional","affiliation":[{"name":"Korea University,School of Electrical Engineering,Seoul,South Korea,02841"}]}],"member":"263","reference":[{"key":"ref13","first-page":"4448","article-title":"Sparsified SGD with memory","volume":"31","author":"stich","year":"2018","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3510587"},{"key":"ref15","article-title":"Understanding top-k sparsification in distributed deep learning","author":"shi","year":"2019"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1045"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2020.3046971"},{"key":"ref10","article-title":"Distributed learning with compressed gradient differences","author":"mishchenko","year":"2019"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-7908-2604-3_16"},{"key":"ref1","first-page":"2595","article-title":"Parallelized stochastic gradient descent","volume":"23","author":"zinkevich","year":"2010","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref17","first-page":"1299","article-title":"Gradient sparsification for communication-efficient distributed optimization","volume":"31","author":"wangni","year":"2018","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref16","first-page":"5973","article-title":"The convergence of sparsified gradient methods","volume":"31","author":"alistarh","year":"2018","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref19","first-page":"8253","article-title":"Fetchsgd: Communication-efficient federated learning with sketching","author":"rothchild","year":"2020","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref18","article-title":"Deep gradient compression: Reducing the communication bandwidth for distributed training","author":"lin","year":"2017"},{"key":"ref24","article-title":"signsgd with majority vote is communication efficient and fault tolerant","author":"bernstein","year":"2018"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/3295500.3356222"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT54713.2023.10206480"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TPDS.2022.3154387"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2019.2944481"},{"key":"ref21","first-page":"14695","article-title":"Qsparse-local-SGD: Distributed SGD with quantization, sparsification and local computations","volume":"32","author":"basu","year":"2019","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2015-354"},{"key":"ref7","first-page":"1","article-title":"TernGrad: Ternary gradients to reduce communication in distributed deep learning","volume":"30","author":"wen","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref9","first-page":"3252","article-title":"Error feedback fixes signsgd and other gradient compression schemes","author":"karimireddy","year":"2019","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref4","article-title":"Revisiting distributed synchronous SGD","author":"chen","year":"2016"},{"key":"ref3","article-title":"Local SGD converges fast and communicates little","author":"stich","year":"2018"},{"key":"ref6","first-page":"560","article-title":"signSGD: Compressed optimisation for non-convex problems","author":"bernstein","year":"2018","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref5","first-page":"1709","article-title":"QSGD: Communication-efficient SGD via gradient quantization and encoding","volume":"30","author":"alistarh","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"}],"event":{"name":"2023 IEEE International Symposium on Information Theory (ISIT)","location":"Taipei, Taiwan","start":{"date-parts":[[2023,6,25]]},"end":{"date-parts":[[2023,6,30]]}},"container-title":["2023 IEEE International Symposium on Information Theory (ISIT)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10206429\/10206441\/10206480.pdf?arnumber=10206480","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T20:28:08Z","timestamp":1773347288000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10206480\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,25]]},"references-count":26,"URL":"https:\/\/doi.org\/10.1109\/isit54713.2023.10206480","relation":{},"subject":[],"published":{"date-parts":[[2023,6,25]]}}}