{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T15:38:24Z","timestamp":1759333104090,"version":"3.37.3"},"reference-count":19,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T00:00:00Z","timestamp":1652659200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T00:00:00Z","timestamp":1652659200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006190","name":"Research and Development","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006190","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,5,16]]},"DOI":"10.1109\/icc45855.2022.9839126","type":"proceedings-article","created":{"date-parts":[[2022,8,11]],"date-time":"2022-08-11T19:37:11Z","timestamp":1660246631000},"page":"207-212","source":"Crossref","is-referenced-by-count":1,"title":["PipeCompress: Accelerating Pipelined Communication for Distributed Deep Learning"],"prefix":"10.1109","author":[{"given":"Juncai","family":"Liu","sequence":"first","affiliation":[{"name":"Tsinghua University,Institute for Network Sciences and Cyberspace, BNRist,China"}]},{"given":"Jessie Hui","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua University,Institute for Network Sciences and Cyberspace, BNRist,China"}]},{"given":"Chenghao","family":"Rong","sequence":"additional","affiliation":[{"name":"Tsinghua University,Institute for Network Sciences and Cyberspace, BNRist,China"}]},{"given":"Jilong","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua University,Institute for Network Sciences and Cyberspace, BNRist,China"}]}],"member":"263","reference":[{"key":"ref10","first-page":"181","article-title":"Poseidon: An efficient communication architecture for distributed deep learning on GPU clusters","author":"zhang","year":"2017","journal-title":"USENIX Annual Technical Conference"},{"key":"ref11","article-title":"Priority-based parameter propagation for distributed dnn training","author":"jayarajan","year":"2019","journal-title":"2nd Conference on Systems and Machine Learning (SysML) SysML"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3341301.3359642"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2019.8737367"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM41043.2020.9155446"},{"year":"2020","key":"ref15","article-title":"Cuda multi-process service"},{"key":"ref16","first-page":"947","article-title":"Analysis of large-scale multi-tenant GPU clusters for DNN training workloads","author":"jeon","year":"2019","journal-title":"2019 USENIX Annual Technical Conference (USENIX ATC 19)"},{"key":"ref17","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"International Conference on Learning Representations"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref4","article-title":"QSGD: Communication-efficient sgd via gradient quantization and encoding","author":"alistarh","year":"2017","journal-title":"Advances in neural information processing systems"},{"article-title":"Layer-wise adaptive gradient sparsification for distributed deep learning with convergence guarantees","year":"2019","author":"shi","key":"ref3"},{"key":"ref6","article-title":"1-bit stochastic gradient descent and its application to data-parallel distributed training of speech dnns","author":"seide","year":"2014","journal-title":"Fifteenth Annual Conference of the International Speech Communication Association"},{"key":"ref5","article-title":"Terngrad: Ternary gradients to reduce communication in distributed deep learning","author":"wen","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref8","article-title":"The convergence of sparsified gradient methods","author":"alistarh","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref7","article-title":"Sparsified SGD with memory","author":"stich","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM41043.2020.9155269"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3363554"},{"key":"ref9","article-title":"Deep gradient compression: Reducing the communication bandwidth for distributed training","author":"lin","year":"2018","journal-title":"International Conference on Learning Representations"}],"event":{"name":"ICC 2022 - IEEE International Conference on Communications","start":{"date-parts":[[2022,5,16]]},"location":"Seoul, Korea, Republic of","end":{"date-parts":[[2022,5,20]]}},"container-title":["ICC 2022 - IEEE International Conference on Communications"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9837954\/9838246\/09839126.pdf?arnumber=9839126","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,12]],"date-time":"2022-09-12T20:05:56Z","timestamp":1663013156000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9839126\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,16]]},"references-count":19,"URL":"https:\/\/doi.org\/10.1109\/icc45855.2022.9839126","relation":{},"subject":[],"published":{"date-parts":[[2022,5,16]]}}}