{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:14:03Z","timestamp":1750220043853,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":57,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,1,4]],"date-time":"2023-01-04T00:00:00Z","timestamp":1672790400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100006754","name":"Army Research Laboratory","doi-asserted-by":"publisher","award":["W911NF-17-2-0196"],"award-info":[{"award-number":["W911NF-17-2-0196"]}],"id":[{"id":"10.13039\/100006754","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,1,4]]},"DOI":"10.1145\/3571306.3571393","type":"proceedings-article","created":{"date-parts":[[2023,1,3]],"date-time":"2023-01-03T16:17:12Z","timestamp":1672762632000},"page":"80-89","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Impact of Redundancy on Resilience in Distributed Optimization and Learning"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4528-3847","authenticated-orcid":false,"given":"Shuo","family":"Liu","sequence":"first","affiliation":[{"name":"Georgetown University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4252-9319","authenticated-orcid":false,"given":"Nirupam","family":"Gupta","sequence":"additional","affiliation":[{"name":"\u00c9cole Polytechnique F\u00e9d\u00e9rale de Lausanne, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5104-8977","authenticated-orcid":false,"given":"Nitin H.","family":"Vaidya","sequence":"additional","affiliation":[{"name":"Georgetown University, United States"}]}],"member":"320","published-online":{"date-parts":[[2023,1,4]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2012.6426626"},{"key":"e_1_3_2_1_2_1","unstructured":"Zeyuan Allen-Zhu and Elad Hazan. 2016. Optimal black-box reductions between optimization objectives. arXiv preprint arXiv:1603.05642(2016).  Zeyuan Allen-Zhu and Elad Hazan. 2016. Optimal black-box reductions between optimization objectives. arXiv preprint arXiv:1603.05642(2016)."},{"key":"e_1_3_2_1_3_1","volume-title":"10th USENIX Symposium on Networked Systems Design and Implementation (NSDI 13)","author":"Ananthanarayanan Ganesh","year":"2013","unstructured":"Ganesh Ananthanarayanan , Ali Ghodsi , Scott Shenker , and Ion Stoica . 2013 . Effective straggler mitigation: Attack of the clones . In 10th USENIX Symposium on Networked Systems Design and Implementation (NSDI 13) . 185\u2013198. Ganesh Ananthanarayanan, Ali Ghodsi, Scott Shenker, and Ion Stoica. 2013. Effective straggler mitigation: Attack of the clones. In 10th USENIX Symposium on Networked Systems Design and Implementation (NSDI 13). 185\u2013198."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2020.3026619"},{"key":"e_1_3_2_1_5_1","volume-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems. 118\u2013128","author":"Blanchard Peva","year":"2017","unstructured":"Peva Blanchard , El\u00a0Mahdi El\u00a0Mhamdi , Rachid Guerraoui , and Julien Stainer . 2017 . Machine learning with adversaries: Byzantine tolerant gradient descent . In Proceedings of the 31st International Conference on Neural Information Processing Systems. 118\u2013128 . Peva Blanchard, El\u00a0Mahdi El\u00a0Mhamdi, Rachid Guerraoui, and Julien Stainer. 2017. Machine learning with adversaries: Byzantine tolerant gradient descent. In Proceedings of the 31st International Conference on Neural Information Processing Systems. 118\u2013128."},{"key":"e_1_3_2_1_6_1","volume-title":"Online learning and stochastic approximations. On-line learning in neural networks 17, 9","author":"Bottou L\u00e9on","year":"1998","unstructured":"L\u00e9on Bottou . 1998. Online learning and stochastic approximations. On-line learning in neural networks 17, 9 ( 1998 ), 142. L\u00e9on Bottou. 1998. Online learning and stochastic approximations. On-line learning in neural networks 17, 9 (1998), 142."},{"key":"e_1_3_2_1_7_1","volume-title":"Advances in Neural Information Processing Systems 20 (NIPS","author":"Bottou L\u00e9on","year":"2007","unstructured":"L\u00e9on Bottou and Olivier Bousquet . 2008. The Tradeoffs of Large Scale Learning . In Advances in Neural Information Processing Systems 20 (NIPS 2007 ). NIPS Foundation , 161\u2013168. L\u00e9on Bottou and Olivier Bousquet. 2008. The Tradeoffs of Large Scale Learning. In Advances in Neural Information Processing Systems 20 (NIPS 2007). NIPS Foundation, 161\u2013168."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1137\/16M1080173"},{"volume-title":"Distributed optimization and statistical learning via the alternating direction method of multipliers","author":"Boyd Stephen","key":"e_1_3_2_1_9_1","unstructured":"Stephen Boyd , Neal Parikh , and Eric Chu . 2011. Distributed optimization and statistical learning via the alternating direction method of multipliers . Now Publishers Inc . Stephen Boyd, Neal Parikh, and Eric Chu. 2011. Distributed optimization and statistical learning via the alternating direction method of multipliers. Now Publishers Inc."},{"key":"e_1_3_2_1_10_1","volume-title":"International Conference on Machine Learning. PMLR, 903\u2013912","author":"Chen Lingjiao","year":"2018","unstructured":"Lingjiao Chen , Hongyi Wang , Zachary Charles , and Dimitris Papailiopoulos . 2018 . Draco: Byzantine-resilient distributed training via redundant gradients . In International Conference on Machine Learning. PMLR, 903\u2013912 . Lingjiao Chen, Hongyi Wang, Zachary Charles, and Dimitris Papailiopoulos. 2018. Draco: Byzantine-resilient distributed training via redundant gradients. In International Conference on Machine Learning. PMLR, 903\u2013912."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3154503","article-title":"Distributed statistical machine learning in adversarial settings: Byzantine gradient descent","volume":"1","author":"Chen Yudong","year":"2017","unstructured":"Yudong Chen , Lili Su , and Jiaming Xu . 2017 . Distributed statistical machine learning in adversarial settings: Byzantine gradient descent . Proceedings of the ACM on Measurement and Analysis of Computing Systems 1 , 2(2017), 1 \u2013 25 . Yudong Chen, Lili Su, and Jiaming Xu. 2017. Distributed statistical machine learning in adversarial settings: Byzantine gradient descent. Proceedings of the ACM on Measurement and Analysis of Computing Systems 1, 2(2017), 1\u201325.","journal-title":"Proceedings of the ACM on Measurement and Analysis of Computing Systems"},{"key":"e_1_3_2_1_12_1","unstructured":"Tarin Clanuwat Mikel Bober-Irizar Asanobu Kitamoto Alex Lamb Kazuaki Yamamoto and David Ha. 2018. Deep Learning for Classical Japanese Literature. arXiv preprint arXiv:1812.01718(2018).  Tarin Clanuwat Mikel Bober-Irizar Asanobu Kitamoto Alex Lamb Kazuaki Yamamoto and David Ha. 2018. Deep Learning for Classical Japanese Literature. arXiv preprint arXiv:1812.01718(2018)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.advwatres.2011.04.013"},{"key":"e_1_3_2_1_14_1","volume-title":"The Conference on Systems and Machine Learning (SysML)","author":"Damaskinos Georgios","year":"2019","unstructured":"Georgios Damaskinos , El\u00a0Mahdi El\u00a0Mhamdi , Rachid Guerraoui , Arsany Hany\u00a0Abdelmessih Guirguis , and S\u00e9bastien Louis\u00a0Alexandre Rouault . 2019 . Aggregathor: Byzantine machine learning via robust gradient aggregation . In The Conference on Systems and Machine Learning (SysML) , 2019. Georgios Damaskinos, El\u00a0Mahdi El\u00a0Mhamdi, Rachid Guerraoui, Arsany Hany\u00a0Abdelmessih Guirguis, and S\u00e9bastien Louis\u00a0Alexandre Rouault. 2019. Aggregathor: Byzantine machine learning via robust gradient aggregation. In The Conference on Systems and Machine Learning (SysML), 2019."},{"key":"e_1_3_2_1_15_1","volume-title":"SAGA: A fast incremental gradient method with support for non-strongly convex composite objectives. arXiv preprint arXiv:1407.0202(2014).","author":"Defazio Aaron","year":"2014","unstructured":"Aaron Defazio , Francis Bach , and Simon Lacoste-Julien . 2014 . SAGA: A fast incremental gradient method with support for non-strongly convex composite objectives. arXiv preprint arXiv:1407.0202(2014). Aaron Defazio, Francis Bach, and Simon Lacoste-Julien. 2014. SAGA: A fast incremental gradient method with support for non-strongly convex composite objectives. arXiv preprint arXiv:1407.0202(2014)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2016.2525015"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2796314.2745873"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"Nirupam Gupta Shuo Liu and Nitin Vaidya. 2021. Byzantine Fault-Tolerant Distributed Machine Learning with Norm-Based Comparative Gradient Elimination. In 2021 51st Annual IEEE\/IFIP International Conference on Dependable Systems and Networks Workshops (DSN-W). IEEE 175\u2013181.  Nirupam Gupta Shuo Liu and Nitin Vaidya. 2021. Byzantine Fault-Tolerant Distributed Machine Learning with Norm-Based Comparative Gradient Elimination. In 2021 51st Annual IEEE\/IFIP International Conference on Dependable Systems and Networks Workshops (DSN-W). IEEE 175\u2013181.","DOI":"10.1109\/DSN-W52860.2021.00037"},{"key":"e_1_3_2_1_19_1","unstructured":"Nirupam Gupta and Nitin\u00a0H Vaidya. 2020. Resilience in collaborative optimization: redundant and independent cost functions. arXiv preprint arXiv:2003.09675(2020).  Nirupam Gupta and Nitin\u00a0H Vaidya. 2020. Resilience in collaborative optimization: redundant and independent cost functions. arXiv preprint arXiv:2003.09675(2020)."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT.2018.8437467"},{"key":"e_1_3_2_1_21_1","unstructured":"Robert Hannah and Wotao Yin. 2017. More iterations per second same quality\u2013why asynchronous algorithms may drastically outperform traditional ones. arXiv preprint arXiv:1708.05136(2017).  Robert Hannah and Wotao Yin. 2017. More iterations per second same quality\u2013why asynchronous algorithms may drastically outperform traditional ones. arXiv preprint arXiv:1708.05136(2017)."},{"key":"e_1_3_2_1_22_1","volume-title":"Accelerating stochastic gradient descent using predictive variance reduction. Advances in neural information processing systems 26","author":"Johnson Rie","year":"2013","unstructured":"Rie Johnson and Tong Zhang . 2013. Accelerating stochastic gradient descent using predictive variance reduction. Advances in neural information processing systems 26 ( 2013 ), 315\u2013323. Rie Johnson and Tong Zhang. 2013. Accelerating stochastic gradient descent using predictive variance reduction. Advances in neural information processing systems 26 (2013), 315\u2013323."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT.2017.8007058"},{"key":"e_1_3_2_1_24_1","first-page":"5434","article-title":"Straggler mitigation in distributed optimization through data encoding","volume":"30","author":"Karakus Can","year":"2017","unstructured":"Can Karakus , Yifan Sun , Suhas Diggavi , and Wotao Yin . 2017 . Straggler mitigation in distributed optimization through data encoding . Advances in Neural Information Processing Systems 30 (2017), 5434 \u2013 5442 . Can Karakus, Yifan Sun, Suhas Diggavi, and Wotao Yin. 2017. Straggler mitigation in distributed optimization through data encoding. Advances in Neural Information Processing Systems 30 (2017), 5434\u20135442.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.5555\/3322706.3362013"},{"key":"e_1_3_2_1_26_1","volume-title":"International Conference on Machine Learning. PMLR, 5311\u20135319","author":"Karimireddy Sai\u00a0Praneeth","year":"2021","unstructured":"Sai\u00a0Praneeth Karimireddy , Lie He , and Martin Jaggi . 2021 . Learning from history for byzantine robust optimization . In International Conference on Machine Learning. PMLR, 5311\u20135319 . Sai\u00a0Praneeth Karimireddy, Lie He, and Martin Jaggi. 2021. Learning from history for byzantine robust optimization. In International Conference on Machine Learning. PMLR, 5311\u20135319."},{"key":"e_1_3_2_1_27_1","unstructured":"Jakub Kone\u010dn\u1ef3 Brendan McMahan and Daniel Ramage. 2015. Federated optimization: Distributed optimization beyond the datacenter. arXiv preprint arXiv:1511.03575(2015).  Jakub Kone\u010dn\u1ef3 Brendan McMahan and Daniel Ramage. 2015. Federated optimization: Distributed optimization beyond the datacenter. arXiv preprint arXiv:1511.03575(2015)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/357172.357176"},{"key":"e_1_3_2_1_29_1","unstructured":"John Langford Alexander Smola and Martin Zinkevich. 2009. Slow learners are fast. arXiv preprint arXiv:0911.0491(2009).  John Langford Alexander Smola and Martin Zinkevich. 2009. Slow learners are fast. arXiv preprint arXiv:0911.0491(2009)."},{"volume-title":"Asynchronous optimization for machine learning. Ph.\u00a0D. Dissertation","author":"Leblond R\u00e9mi","key":"e_1_3_2_1_30_1","unstructured":"R\u00e9mi Leblond . 2018. Asynchronous optimization for machine learning. Ph.\u00a0D. Dissertation . PSL Research University . R\u00e9mi Leblond. 2018. Asynchronous optimization for machine learning. Ph.\u00a0D. Dissertation. PSL Research University."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2017.2736066"},{"key":"e_1_3_2_1_33_1","first-page":"19","article-title":"Communication efficient distributed machine learning with the parameter server","volume":"27","author":"Li Mu","year":"2014","unstructured":"Mu Li , David\u00a0 G Andersen , Alexander\u00a0 J Smola , and Kai Yu . 2014 . Communication efficient distributed machine learning with the parameter server . Advances in Neural Information Processing Systems 27 (2014), 19 \u2013 27 . Mu Li, David\u00a0G Andersen, Alexander\u00a0J Smola, and Kai Yu. 2014. Communication efficient distributed machine learning with the parameter server. Advances in Neural Information Processing Systems 27 (2014), 19\u201327.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_34_1","unstructured":"Shuo Liu. 2021. A Survey on Fault-tolerance in Distributed Optimization and Machine Learning. arXiv preprint arXiv:2106.08545(2021).  Shuo Liu. 2021. A Survey on Fault-tolerance in Distributed Optimization and Machine Learning. arXiv preprint arXiv:2106.08545(2021)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3465084.3467902"},{"key":"e_1_3_2_1_36_1","unstructured":"Shuo Liu Nirupam Gupta and Nitin\u00a0H. Vaidya. 2022. Impact of Redundancy on Resilience in Distributed Optimization and Learning. arXiv preprint arXiv:2211.08622(2022).  Shuo Liu Nirupam Gupta and Nitin\u00a0H. Vaidya. 2022. Impact of Redundancy on Resilience in Distributed Optimization and Learning. arXiv preprint arXiv:2211.08622(2022)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2008.2009515"},{"key":"e_1_3_2_1_38_1","unstructured":"Feng Niu Benjamin Recht Christopher R\u00e9 and Stephen\u00a0J Wright. 2011. Hogwild!: A lock-free approach to parallelizing stochastic gradient descent. arXiv preprint arXiv:1106.5730(2011).  Feng Niu Benjamin Recht Christopher R\u00e9 and Stephen\u00a0J Wright. 2011. Hogwild!: A lock-free approach to parallelizing stochastic gradient descent. arXiv preprint arXiv:1106.5730(2011)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.2979670"},{"key":"e_1_3_2_1_40_1","volume-title":"Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703(2019).","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , Luca Antiga , 2019 . Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703(2019). Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, 2019. Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703(2019)."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/984622.984626"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2004.1428778"},{"key":"e_1_3_2_1_43_1","unstructured":"Nicolas\u00a0Le Roux Mark Schmidt and Francis Bach. 2012. A stochastic gradient method with an exponential convergence rate for finite training sets. arXiv preprint arXiv:1202.6258(2012).  Nicolas\u00a0Le Roux Mark Schmidt and Francis Bach. 2012. A stochastic gradient method with an exponential convergence rate for finite training sets. arXiv preprint arXiv:1202.6258(2012)."},{"key":"e_1_3_2_1_44_1","volume-title":"When do redundant requests reduce latency?IEEE Transactions on Communications 64, 2","author":"Shah B","year":"2015","unstructured":"Nihar\u00a0 B Shah , Kangwook Lee , and Kannan Ramchandran . 2015. When do redundant requests reduce latency?IEEE Transactions on Communications 64, 2 ( 2015 ), 715\u2013722. Nihar\u00a0B Shah, Kangwook Lee, and Kannan Ramchandran. 2015. When do redundant requests reduce latency?IEEE Transactions on Communications 64, 2 (2015), 715\u2013722."},{"key":"e_1_3_2_1_45_1","unstructured":"Shai Shalev-Shwartz and Tong Zhang. 2013. Accelerated mini-batch stochastic dual coordinate ascent. arXiv preprint arXiv:1305.2581(2013).  Shai Shalev-Shwartz and Tong Zhang. 2013. Accelerated mini-batch stochastic dual coordinate ascent. arXiv preprint arXiv:1305.2581(2013)."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1137\/14096668X"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"crossref","unstructured":"S. Sra S. Nowozin and S.J. Wright. 2012. Optimization for Machine Learning. MIT Press.  S. Sra S. Nowozin and S.J. Wright. 2012. Optimization for Machine Learning. MIT Press.","DOI":"10.7551\/mitpress\/8996.001.0001"},{"key":"e_1_3_2_1_48_1","unstructured":"Lili Su and Nitin Vaidya. 2015. Byzantine multi-agent optimization: Part I. arXiv preprint arXiv:1506.04681(2015).  Lili Su and Nitin Vaidya. 2015. Byzantine multi-agent optimization: Part I. arXiv preprint arXiv:1506.04681(2015)."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/2933057.2933105"},{"key":"e_1_3_2_1_50_1","volume-title":"International Conference on Machine Learning. PMLR, 3368\u20133376","author":"Tandon Rashish","year":"2017","unstructured":"Rashish Tandon , Qi Lei , Alexandros\u00a0 G Dimakis , and Nikos Karampatziakis . 2017 . Gradient coding: Avoiding stragglers in distributed learning . In International Conference on Machine Learning. PMLR, 3368\u20133376 . Rashish Tandon, Qi Lei, Alexandros\u00a0G Dimakis, and Nikos Karampatziakis. 2017. Gradient coding: Avoiding stragglers in distributed learning. In International Conference on Machine Learning. PMLR, 3368\u20133376."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2015.2449811"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/2847220.2847223"},{"key":"e_1_3_2_1_53_1","unstructured":"Han Xiao Kashif Rasul and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747(2017).  Han Xiao Kashif Rasul and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747(2017)."},{"key":"e_1_3_2_1_54_1","volume-title":"Zeno: Byzantine-suspicious stochastic gradient descent. arXiv preprint arXiv:1805.10032 24","author":"Xie Cong","year":"2018","unstructured":"Cong Xie , Oluwasanmi Koyejo , and Indranil Gupta . 2018 . Zeno: Byzantine-suspicious stochastic gradient descent. arXiv preprint arXiv:1805.10032 24 (2018). Cong Xie, Oluwasanmi Koyejo, and Indranil Gupta. 2018. Zeno: Byzantine-suspicious stochastic gradient descent. arXiv preprint arXiv:1805.10032 24 (2018)."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.5555\/2946645.3007059"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.5555\/3294771.3294839"},{"key":"e_1_3_2_1_57_1","volume-title":"International Conference on Machine Learning. PMLR, 5650\u20135659","author":"Yin Dong","year":"2018","unstructured":"Dong Yin , Yudong Chen , Ramchandran Kannan , and Peter Bartlett . 2018 . Byzantine-robust distributed learning: Towards optimal statistical rates . In International Conference on Machine Learning. PMLR, 5650\u20135659 . Dong Yin, Yudong Chen, Ramchandran Kannan, and Peter Bartlett. 2018. Byzantine-robust distributed learning: Towards optimal statistical rates. In International Conference on Machine Learning. PMLR, 5650\u20135659."}],"event":{"name":"ICDCN 2023: 24th International Conference on Distributed Computing and Networking","acronym":"ICDCN 2023","location":"Kharagpur India"},"container-title":["Proceedings of the 24th International Conference on Distributed Computing and Networking"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3571306.3571393","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3571306.3571393","content-type":"text\/html","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3571306.3571393","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3571306.3571393","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:07:53Z","timestamp":1750183673000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3571306.3571393"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,4]]},"references-count":57,"alternative-id":["10.1145\/3571306.3571393","10.1145\/3571306"],"URL":"https:\/\/doi.org\/10.1145\/3571306.3571393","relation":{},"subject":[],"published":{"date-parts":[[2023,1,4]]},"assertion":[{"value":"2023-01-04","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}