{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,9]],"date-time":"2026-01-09T17:47:58Z","timestamp":1767980878495,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,6,9]],"date-time":"2021-06-09T00:00:00Z","timestamp":1623196800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,6,9]]},"DOI":"10.1145\/3448016.3452773","type":"proceedings-article","created":{"date-parts":[[2021,6,18]],"date-time":"2021-06-18T17:22:37Z","timestamp":1624036957000},"page":"2262-2270","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":58,"title":["Heterogeneity-Aware Distributed Machine Learning Training via Partial Reduce"],"prefix":"10.1145","author":[{"given":"Xupeng","family":"Miao","sequence":"first","affiliation":[{"name":"Peking University &amp; Tencent Inc., Beijing, China"}]},{"given":"Xiaonan","family":"Nie","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"given":"Yingxia","family":"Shao","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications, Beijing, China"}]},{"given":"Zhi","family":"Yang","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"given":"Jiawei","family":"Jiang","sequence":"additional","affiliation":[{"name":"ETH Z\u00fcrich, Zurich, Switzerland"}]},{"given":"Lingxiao","family":"Ma","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"given":"Bin","family":"Cui","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2021,6,18]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Benoit Steiner, Paul A. Tucker, Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng.","author":"Abadi Mart'i","year":"2016","unstructured":"Mart'i n Abadi , Paul Barham , Jianmin Chen , Zhifeng Chen , Andy Davis , Jeffrey Dean , Matthieu Devin , Sanjay Ghemawat , Geoffrey Irving , Michael Isard , Manjunath Kudlur , Josh Levenberg , Rajat Monga , Sherry Moore , Derek Gordon Murray , Benoit Steiner, Paul A. Tucker, Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. 2016 . TensorFlow: A System for Large-Scale Machine Learning. In OSDI. 265--283. Mart'i n Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, Manjunath Kudlur, Josh Levenberg, Rajat Monga, Sherry Moore, Derek Gordon Murray, Benoit Steiner, Paul A. Tucker, Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. 2016. TensorFlow: A System for Large-Scale Machine Learning. In OSDI. 265--283."},{"key":"e_1_3_2_2_2_1","first-page":"61","article-title":"Novel Parallel Algorithms for Fast Multi-GPU-Based Generation of Massive Scale-Free Networks. Data Sci","volume":"4","author":"Maksudul Alam Md.","year":"2019","unstructured":"Md. Maksudul Alam , Kalyan S. Perumalla , and Peter Sanders . 2019 . Novel Parallel Algorithms for Fast Multi-GPU-Based Generation of Massive Scale-Free Networks. Data Sci . Eng. , Vol. 4 , 1 (2019), 61 -- 75 . Md. Maksudul Alam, Kalyan S. Perumalla, and Peter Sanders. 2019. Novel Parallel Algorithms for Fast Multi-GPU-Based Generation of Massive Scale-Free Networks. Data Sci. Eng. , Vol. 4, 1 (2019), 61--75.","journal-title":"Eng."},{"key":"e_1_3_2_2_3_1","volume-title":"QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding. In NeurIPS .","author":"Alistarh Dan","year":"2017","unstructured":"Dan Alistarh , Demjan Grubic , Jerry Li , Ryota Tomioka , and M. Vojnovic . 2017 . QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding. In NeurIPS . Dan Alistarh, Demjan Grubic, Jerry Li, Ryota Tomioka, and M. Vojnovic. 2017. QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding. In NeurIPS ."},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1137\/16M1080173"},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/2898442.2898444"},{"key":"e_1_3_2_2_6_1","unstructured":"J. Chen Rajat Monga S. Bengio and R. J\u00f3zefowicz. 2016. Revisiting Distributed Synchronous SGD. ArXiv Vol. abs\/1702.05800 (2016).  J. Chen Rajat Monga S. Bengio and R. J\u00f3zefowicz. 2016. Revisiting Distributed Synchronous SGD. ArXiv Vol. abs\/1702.05800 (2016)."},{"key":"e_1_3_2_2_7_1","volume-title":"Ubershuffle: Communication-efficient data shuffling for sgd via coding theory. NeurIPS .","author":"Chung Jichan","year":"2017","unstructured":"Jichan Chung , Kangwook Lee , Ramtin Pedarsani , Dimitris Papailiopoulos , and Kannan Ramchandran . 2017 . Ubershuffle: Communication-efficient data shuffling for sgd via coding theory. NeurIPS . Jichan Chung, Kangwook Lee, Ramtin Pedarsani, Dimitris Papailiopoulos, and Kannan Ramchandran. 2017. Ubershuffle: Communication-efficient data shuffling for sgd via coding theory. NeurIPS ."},{"key":"e_1_3_2_2_8_1","volume-title":"Ng","author":"Dean Jeffrey","year":"2012","unstructured":"Jeffrey Dean , Greg Corrado , Rajat Monga , Kai Chen , Matthieu Devin , Quoc V. Le , Mark Z. Mao , Marc'Aurelio Ranzato , Andrew W. Senior , Paul A. Tucker , Ke Yang , and Andrew Y . Ng . 2012 . Large Scale Distributed Deep Networks. In NeurIPS. 1232--1240. Jeffrey Dean, Greg Corrado, Rajat Monga, Kai Chen, Matthieu Devin, Quoc V. Le, Mark Z. Mao, Marc'Aurelio Ranzato, Andrew W. Senior, Paul A. Tucker, Ke Yang, and Andrew Y. Ng. 2012. Large Scale Distributed Deep Networks. In NeurIPS. 1232--1240."},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"crossref","unstructured":"Jia Deng Wei Dong Richard Socher Li-Jia Li Kai Li and Fei-Fei Li. 2009. ImageNet: A large-scale hierarchical image database. In CVPR. 248--255.  Jia Deng Wei Dong Richard Socher Li-Jia Li Kai Li and Fei-Fei Li. 2009. ImageNet: A large-scale hierarchical image database. In CVPR. 248--255.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_2_10_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171--4186.","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171--4186. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171--4186."},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11432-020-2952-7"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"crossref","unstructured":"Arash Fard Anh Le George Larionov Waqas Dhillon and Chuck Bear. 2020. Vertica-ML: Distributed Machine Learning in Vertica Database. In SIGMOD. 755--768.  Arash Fard Anh Le George Larionov Waqas Dhillon and Chuck Bear. 2020. Vertica-ML: Distributed Machine Learning in Vertica Database. In SIGMOD. 755--768.","DOI":"10.1145\/3318464.3386137"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10107-014-0846-1"},{"key":"e_1_3_2_2_14_1","unstructured":"Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In CVPR. 770--778.  Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In CVPR. 770--778."},{"key":"e_1_3_2_2_15_1","volume-title":"Phillip B. Gibbons, Garth A. Gibson, Gregory R. Ganger, and Eric P. Xing.","author":"Ho Qirong","year":"2013","unstructured":"Qirong Ho , James Cipar , Henggang Cui , Seunghak Lee , Jin Kyu Kim , Phillip B. Gibbons, Garth A. Gibson, Gregory R. Ganger, and Eric P. Xing. 2013 . More Effective Distributed ML via a Stale Synchronous Parallel Parameter Server. In NeurIPS. 1223--1231. Qirong Ho, James Cipar, Henggang Cui, Seunghak Lee, Jin Kyu Kim, Phillip B. Gibbons, Garth A. Gibson, Gregory R. Ganger, and Eric P. Xing. 2013. More Effective Distributed ML via a Stale Synchronous Parallel Parameter Server. In NeurIPS. 1223--1231."},{"key":"e_1_3_2_2_16_1","volume-title":"Gaia: Geo-Distributed Machine Learning Approaching LAN Speeds. In NSDI . 629--647.","author":"Hsieh Kevin","year":"2017","unstructured":"Kevin Hsieh , Aaron Harlap , Nandita Vijaykumar , Dimitris Konomis , Gregory R. Ganger , Phillip B. Gibbons , and Onur Mutlu . 2017 . Gaia: Geo-Distributed Machine Learning Approaching LAN Speeds. In NSDI . 629--647. Kevin Hsieh, Aaron Harlap, Nandita Vijaykumar, Dimitris Konomis, Gregory R. Ganger, Phillip B. Gibbons, and Onur Mutlu. 2017. Gaia: Geo-Distributed Machine Learning Approaching LAN Speeds. In NSDI . 629--647."},{"key":"e_1_3_2_2_17_1","volume-title":"Weinberger","author":"Huang Gao","year":"2017","unstructured":"Gao Huang , Zhuang Liu , Laurens van der Maaten , and Kilian Q . Weinberger . 2017 . Densely Connected Convolutional Networks. In CVPR. 2261--2269. Gao Huang, Zhuang Liu, Laurens van der Maaten, and Kilian Q. Weinberger. 2017. Densely Connected Convolutional Networks. In CVPR. 2261--2269."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"crossref","unstructured":"Jiawei Jiang Bin Cui Ce Zhang and Lele Yu. 2017. Heterogeneity-aware Distributed Parameter Servers. In SIGMOD. 463--478.  Jiawei Jiang Bin Cui Ce Zhang and Lele Yu. 2017. Heterogeneity-aware Distributed Parameter Servers. In SIGMOD. 463--478.","DOI":"10.1145\/3035918.3035933"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"crossref","unstructured":"Jiawei Jiang Fangcheng Fu Tong Yang and Bin Cui. 2018. SketchML: Accelerating Distributed Machine Learning with Data Sketches. In SIGMOD . 1269--1284.  Jiawei Jiang Fangcheng Fu Tong Yang and Bin Cui. 2018. SketchML: Accelerating Distributed Machine Learning with Data Sketches. In SIGMOD . 1269--1284.","DOI":"10.1145\/3183713.3196894"},{"key":"e_1_3_2_2_20_1","first-page":"1","article-title":"Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks","volume":"43","author":"Kim Soojeong","year":"2019","unstructured":"Soojeong Kim , Gyeong-In Yu , Hojin Park , Sungwoo Cho , Eunji Jeong , Hyeonmin Ha , Sanha Lee , Joo Seong Jeong , and Byung-Gon Chun . 2019 . Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks . In EuroSys . 43 : 1 -- 43 :15. Soojeong Kim, Gyeong-In Yu, Hojin Park, Sungwoo Cho, Eunji Jeong, Hyeonmin Ha, Sanha Lee, Joo Seong Jeong, and Byung-Gon Chun. 2019. Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks. In EuroSys . 43:1--43:15.","journal-title":"EuroSys ."},{"key":"e_1_3_2_2_21_1","first-page":"157","article-title":"Scaling Word2Vec on Big Corpus. Data Sci","volume":"4","author":"Li Bofang","year":"2019","unstructured":"Bofang Li , Aleksandr Drozd , Yuhe Guo , Tao Liu , Satoshi Matsuoka , and Xiaoyong Du . 2019 . Scaling Word2Vec on Big Corpus. Data Sci . Eng. , Vol. 4 , 2 (2019), 157 -- 175 . Bofang Li, Aleksandr Drozd, Yuhe Guo, Tao Liu, Satoshi Matsuoka, and Xiaoyong Du. 2019. Scaling Word2Vec on Big Corpus. Data Sci. Eng. , Vol. 4, 2 (2019), 157--175.","journal-title":"Eng."},{"key":"e_1_3_2_2_22_1","volume-title":"Alexander J. Smola, Amr Ahmed, Vanja Josifovski, James Long, Eugene J. Shekita, and Bor-Yiing Su.","author":"Li Mu","year":"2014","unstructured":"Mu Li , David G. Andersen , Jun Woo Park , Alexander J. Smola, Amr Ahmed, Vanja Josifovski, James Long, Eugene J. Shekita, and Bor-Yiing Su. 2014 . Scaling Distributed Machine Learning with the Parameter Server. In OSDI . 583--598. Mu Li, David G. Andersen, Jun Woo Park, Alexander J. Smola, Amr Ahmed, Vanja Josifovski, James Long, Eugene J. Shekita, and Bor-Yiing Su. 2014. Scaling Distributed Machine Learning with the Parameter Server. In OSDI . 583--598."},{"key":"e_1_3_2_2_23_1","unstructured":"Shigang Li Tal Ben-Nun Salvatore Di Girolamo Dan Alistarh and Torsten Hoefler. 2020 a. Taming unbalanced training workloads in deep learning with partial collective operations. In PPoPP. 45--61.  Shigang Li Tal Ben-Nun Salvatore Di Girolamo Dan Alistarh and Torsten Hoefler. 2020 a. Taming unbalanced training workloads in deep learning with partial collective operations. In PPoPP. 45--61."},{"key":"e_1_3_2_2_24_1","first-page":"3005","article-title":"b","volume":"13","author":"Li Shen","year":"2020","unstructured":"Shen Li , Yanli Zhao , Rohan Varma , Omkar Salpekar , Pieter Noordhuis , Teng Li , Adam Paszke , Jeff Smith , Brian Vaughan , Pritam Damania , and Soumith Chintala . 2020 b . PyTorch Distributed: Experiences on Accelerating Data Parallel Training. PVLDB , Vol. 13 , 12 (2020), 3005 -- 3018 . Shen Li, Yanli Zhao, Rohan Varma, Omkar Salpekar, Pieter Noordhuis, Teng Li, Adam Paszke, Jeff Smith, Brian Vaughan, Pritam Damania, and Soumith Chintala. 2020 b. PyTorch Distributed: Experiences on Accelerating Data Parallel Training. PVLDB , Vol. 13, 12 (2020), 3005--3018.","journal-title":"PyTorch Distributed: Experiences on Accelerating Data Parallel Training. PVLDB"},{"key":"e_1_3_2_2_25_1","unstructured":"Xiangru Lian Yijun Huang Yuncheng Li and Ji Liu. 2015. Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization. In NeurIPS . 2737--2745.  Xiangru Lian Yijun Huang Yuncheng Li and Ji Liu. 2015. Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization. In NeurIPS . 2737--2745."},{"key":"e_1_3_2_2_26_1","unstructured":"Xiangru Lian Ce Zhang Huan Zhang Cho-Jui Hsieh Wei Zhang and Ji Liu. 2017. Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent. In NeruIPS . 5330--5340.  Xiangru Lian Ce Zhang Huan Zhang Cho-Jui Hsieh Wei Zhang and Ji Liu. 2017. Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent. In NeruIPS . 5330--5340."},{"key":"e_1_3_2_2_27_1","first-page":"3049","article-title":"Asynchronous Decentralized Parallel Stochastic Gradient Descent","volume":"80","author":"Lian Xiangru","year":"2018","unstructured":"Xiangru Lian , Wei Zhang , Ce Zhang , and Ji Liu . 2018 . Asynchronous Decentralized Parallel Stochastic Gradient Descent . In ICML , Vol. 80. 3049 -- 3058 . Xiangru Lian, Wei Zhang, Ce Zhang, and Ji Liu. 2018. Asynchronous Decentralized Parallel Stochastic Gradient Descent. In ICML , Vol. 80. 3049--3058.","journal-title":"ICML"},{"key":"e_1_3_2_2_28_1","volume-title":"MixML: A Unified Analysis of Weakly Consistent Parallel Learning. CoRR","author":"Lu Yucheng","year":"2020","unstructured":"Yucheng Lu , Jack Nash , and Christopher De Sa. 2020. MixML: A Unified Analysis of Weakly Consistent Parallel Learning. CoRR , Vol. abs\/ 2005 .06706 ( 2020 ). arxiv: 2005.06706 Yucheng Lu, Jack Nash, and Christopher De Sa. 2020. MixML: A Unified Analysis of Weakly Consistent Parallel Learning. CoRR , Vol. abs\/2005.06706 (2020). arxiv: 2005.06706"},{"key":"e_1_3_2_2_29_1","volume-title":"Prague: High-Performance Heterogeneity-Aware Asynchronous Decentralized Training. In ASPLOS . 401--416.","author":"Luo Qinyi","year":"2020","unstructured":"Qinyi Luo , Jiaao He , Youwei Zhuo , and Xuehai Qian . 2020 . Prague: High-Performance Heterogeneity-Aware Asynchronous Decentralized Training. In ASPLOS . 401--416. Qinyi Luo, Jiaao He, Youwei Zhuo, and Xuehai Qian. 2020. Prague: High-Performance Heterogeneity-Aware Asynchronous Decentralized Training. In ASPLOS . 401--416."},{"key":"#cr-split#-e_1_3_2_2_30_1.1","doi-asserted-by":"crossref","unstructured":"X. Miao L. Ma Z. Yang Y. Shao B. Cui L. Yu and J. Jiang. 2020. CuWide: Towards Efficient Flow-based Training for Sparse Wide Models on GPUs. TKDE (2020) 1--1. https:\/\/doi.org\/10.1109\/TKDE.2020.3038109 10.1109\/TKDE.2020.3038109","DOI":"10.1109\/TKDE.2020.3038109"},{"key":"#cr-split#-e_1_3_2_2_30_1.2","doi-asserted-by":"crossref","unstructured":"X. Miao L. Ma Z. Yang Y. Shao B. Cui L. Yu and J. Jiang. 2020. CuWide: Towards Efficient Flow-based Training for Sparse Wide Models on GPUs. TKDE (2020) 1--1. https:\/\/doi.org\/10.1109\/TKDE.2020.3038109","DOI":"10.1109\/ICDE51399.2021.00251"},{"key":"e_1_3_2_2_31_1","volume-title":"Edward Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala.","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , Luca Antiga , Alban Desmaison , Andreas K\u00f6 pf , Edward Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019 . PyTorch: An Imperative Style, High-Performance Deep Learning Library . , Vol. abs\/ 1912 .01703 (2019). Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas K\u00f6 pf, Edward Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. , Vol. abs\/1912.01703 (2019)."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jpdc.2008.09.002"},{"key":"e_1_3_2_2_33_1","volume-title":"Horovod: fast and easy distributed deep learning in TensorFlow. CoRR","author":"Sergeev Alexander","year":"2018","unstructured":"Alexander Sergeev and Mike Del Balso . 2018. Horovod: fast and easy distributed deep learning in TensorFlow. CoRR , Vol. abs\/ 1802 .05799 ( 2018 ). Alexander Sergeev and Mike Del Balso. 2018. Horovod: fast and easy distributed deep learning in TensorFlow. CoRR , Vol. abs\/1802.05799 (2018)."},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ymssp.2017.08.002"},{"key":"e_1_3_2_2_35_1","unstructured":"Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In ICLR .  Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In ICLR ."},{"key":"e_1_3_2_2_36_1","unstructured":"Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In NeurIPS . 1195--1204.  Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In NeurIPS . 1195--1204."},{"key":"e_1_3_2_2_37_1","volume-title":"Cooperative SGD: A Unified Framework for the Design and Analysis of Communication-Efficient SGD Algorithms. In ICML Workshop .","author":"Wang Jianyu","year":"2019","unstructured":"Jianyu Wang and Gauri Joshi . 2019 . Cooperative SGD: A Unified Framework for the Design and Analysis of Communication-Efficient SGD Algorithms. In ICML Workshop . Jianyu Wang and Gauri Joshi. 2019. Cooperative SGD: A Unified Framework for the Design and Analysis of Communication-Efficient SGD Algorithms. In ICML Workshop ."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"crossref","unstructured":"Xintong Wang and Yunfei Feng. 2018. An Ensemble Learning Algorithm for Indoor Localization. In ICCC . 774--778.  Xintong Wang and Yunfei Feng. 2018. An Ensemble Learning Algorithm for Indoor Localization. In ICCC . 774--778.","DOI":"10.1109\/CompComm.2018.8780770"},{"key":"e_1_3_2_2_39_1","volume-title":"Gandiva: Introspective Cluster Scheduling for Deep Learning. In OSDI. 595--610.","author":"Xiao Wencong","year":"2018","unstructured":"Wencong Xiao , Romil Bhardwaj , Ramachandran Ramjee , Muthian Sivathanu , Nipun Kwatra , Zhenhua Han , Pratyush Patel , Xuan Peng , Hanyu Zhao , Quanlu Zhang , Fan Yang , and Lidong Zhou . 2018 . Gandiva: Introspective Cluster Scheduling for Deep Learning. In OSDI. 595--610. Wencong Xiao, Romil Bhardwaj, Ramachandran Ramjee, Muthian Sivathanu, Nipun Kwatra, Zhenhua Han, Pratyush Patel, Xuan Peng, Hanyu Zhao, Quanlu Zhang, Fan Yang, and Lidong Zhou. 2018. Gandiva: Introspective Cluster Scheduling for Deep Learning. In OSDI. 595--610."},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"crossref","unstructured":"Xu Xie Fei Sun Xiaoyong Yang Zhao Yang Jinyang Gao Wenwu Ou and Bin Cui. 2021. Explore User Neighborhood for Real-time E-commerce Recommendation. arxiv: cs.IR\/2103.00442  Xu Xie Fei Sun Xiaoyong Yang Zhao Yang Jinyang Gao Wenwu Ou and Bin Cui. 2021. Explore User Neighborhood for Real-time E-commerce Recommendation. arxiv: cs.IR\/2103.00442","DOI":"10.1109\/ICDE51399.2021.00279"},{"key":"e_1_3_2_2_41_1","first-page":"1283","article-title":"DimmWitted","volume":"7","author":"Zhang Ce","year":"2014","unstructured":"Ce Zhang and Christopher R\u00e9 . 2014 . DimmWitted : A Study of Main-Memory Statistical Analytics. PVLDB , Vol. 7 , 12 (2014), 1283 -- 1294 . Ce Zhang and Christopher R\u00e9. 2014. DimmWitted: A Study of Main-Memory Statistical Analytics. PVLDB , Vol. 7, 12 (2014), 1283--1294.","journal-title":"A Study of Main-Memory Statistical Analytics. PVLDB"},{"key":"e_1_3_2_2_42_1","unstructured":"Zhipeng Zhang Bin Cui Yingxia Shao Lele Yu Jiawei Jiang and Xupeng Miao. 2019. PS2: Parameter Server on Spark. In SIGMOD. 376--388.  Zhipeng Zhang Bin Cui Yingxia Shao Lele Yu Jiawei Jiang and Xupeng Miao. 2019. PS2: Parameter Server on Spark. In SIGMOD. 376--388."},{"key":"e_1_3_2_2_43_1","unstructured":"Shuai Zheng and James T Kwok. 2017. Follow the moving leader in deep learning. In ICML. 4110--4119.  Shuai Zheng and James T Kwok. 2017. Follow the moving leader in deep learning. In ICML. 4110--4119."},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11432-020-2861-0"},{"key":"e_1_3_2_2_45_1","unstructured":"Martin Zinkevich M. Weimer Alex Smola and L. Li. 2010. Parallelized Stochastic Gradient Descent. In NeurIPS .  Martin Zinkevich M. Weimer Alex Smola and L. Li. 2010. Parallelized Stochastic Gradient Descent. In NeurIPS ."}],"event":{"name":"SIGMOD\/PODS '21: International Conference on Management of Data","location":"Virtual Event China","acronym":"SIGMOD\/PODS '21","sponsor":["SIGMOD ACM Special Interest Group on Management of Data"]},"container-title":["Proceedings of the 2021 International Conference on Management of Data"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3448016.3452773","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3448016.3452773","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:49:14Z","timestamp":1750193354000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3448016.3452773"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,9]]},"references-count":46,"alternative-id":["10.1145\/3448016.3452773","10.1145\/3448016"],"URL":"https:\/\/doi.org\/10.1145\/3448016.3452773","relation":{},"subject":[],"published":{"date-parts":[[2021,6,9]]},"assertion":[{"value":"2021-06-18","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}