{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:23:53Z","timestamp":1750220633096,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":29,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,12,1]],"date-time":"2020-12-01T00:00:00Z","timestamp":1606780800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Research Foundation of Korea","award":["2019M3 F2A107223912"],"award-info":[{"award-number":["2019M3 F2A107223912"]}]},{"name":"Institute for Information and Communications Technology Promotion","award":["2018-0-00278"],"award-info":[{"award-number":["2018-0-00278"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2020,12]]},"DOI":"10.1145\/3426745.3431335","type":"proceedings-article","created":{"date-parts":[[2020,11,26]],"date-time":"2020-11-26T20:27:41Z","timestamp":1606422461000},"page":"1-6","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["FEWER"],"prefix":"10.1145","author":[{"given":"Yongjin","family":"Shin","sequence":"first","affiliation":[{"name":"KAIST Industrial&amp;Systems Engineering, Daejeon, Rep. of Korea"}]},{"given":"Gihun","family":"Lee","sequence":"additional","affiliation":[{"name":"KAIST Graduate School of AI, Daejeon, Rep. of Korea"}]},{"given":"Seungjae","family":"Shin","sequence":"additional","affiliation":[{"name":"KAIST Industrial&amp;Systems Engineering, Daejeon, Rep. of Korea"}]},{"given":"Se-young","family":"Yun","sequence":"additional","affiliation":[{"name":"KAIST Graduate School of AI, Daejeon, Rep. of Korea"}]},{"given":"Il-chul","family":"Moon","sequence":"additional","affiliation":[{"name":"KAIST Industrial&amp;Systems Engineering Daejeon, Rep. of Korea"}]}],"member":"320","published-online":{"date-parts":[[2020,12]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3013541"},{"key":"e_1_3_2_2_2_1","volume-title":"A simple method for reconstructing a high-quality NDVI time-series data set based on the Savitzky-Golay filter. Remote sensing of Environment 91, 3-4","author":"Chen Jin","year":"2004","unstructured":"Jin Chen , Per J\u00f6nsson , Masayuki Tamura , Zhihui Gu , Bunkei Matsushita , and Lars Eklundh . 2004. A simple method for reconstructing a high-quality NDVI time-series data set based on the Savitzky-Golay filter. Remote sensing of Environment 91, 3-4 ( 2004 ), 332--344. Jin Chen, Per J\u00f6nsson, Masayuki Tamura, Zhihui Gu, Bunkei Matsushita, and Lars Eklundh. 2004. A simple method for reconstructing a high-quality NDVI time-series data set based on the Savitzky-Golay filter. Remote sensing of Environment 91, 3-4 (2004), 332--344."},{"key":"e_1_3_2_2_3_1","volume-title":"A survey of model compression and acceleration for deep neural networks. arXiv preprint arXiv:1710.09282","author":"Cheng Yu","year":"2017","unstructured":"Yu Cheng , Duo Wang , Pan Zhou , and Tao Zhang . 2017. A survey of model compression and acceleration for deep neural networks. arXiv preprint arXiv:1710.09282 ( 2017 ). Yu Cheng, Duo Wang, Pan Zhou, and Tao Zhang. 2017. A survey of model compression and acceleration for deep neural networks. arXiv preprint arXiv:1710.09282 (2017)."},{"key":"e_1_3_2_2_4_1","volume-title":"The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635","author":"Frankle Jonathan","year":"2018","unstructured":"Jonathan Frankle and Michael Carbin . 2018. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635 ( 2018 ). Jonathan Frankle and Michael Carbin. 2018. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635 (2018)."},{"key":"e_1_3_2_2_5_1","volume-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149","author":"Han Song","year":"2015","unstructured":"Song Han , Huizi Mao , and William J Dally . 2015. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 ( 2015 ). Song Han, Huizi Mao, and William J Dally. 2015. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)."},{"key":"e_1_3_2_2_6_1","unstructured":"Ching-Yi Hung Cheng-Hao Tu Cheng-En Wu Chien-Hung Chen Yi-Ming Chan and Chu-Song Chen. 2019. Compacting Picking and Growing for Unforgetting Continual Learning. In Advances in Neural Information Processing Systems. 13669--13679.  Ching-Yi Hung Cheng-Hao Tu Cheng-En Wu Chien-Hung Chen Yi-Ming Chan and Chu-Song Chen. 2019. Compacting Picking and Growing for Unforgetting Continual Learning. In Advances in Neural Information Processing Systems. 13669--13679."},{"key":"e_1_3_2_2_7_1","volume-title":"SGDR: Stochastic Gradient Descent with Warm Restarts. In International Conference on Learning Representations","author":"Ilya Loshchilov Frank Hutter","year":"2017","unstructured":"Frank Hutter Ilya Loshchilov . 2017 . SGDR: Stochastic Gradient Descent with Warm Restarts. In International Conference on Learning Representations (2017). Frank Hutter Ilya Loshchilov. 2017. SGDR: Stochastic Gradient Descent with Warm Restarts. In International Conference on Learning Representations (2017)."},{"key":"e_1_3_2_2_8_1","volume-title":"Wei-Han Lee, and Leandros Tassiulas.","author":"Jiang Yuang","year":"2020","unstructured":"Yuang Jiang , Shiqiang Wang , Bong Jun Ko , Wei-Han Lee, and Leandros Tassiulas. 2020 . Model Pruning Enables Efficient Federated Learning on Edge Devices . arXiv preprint arXiv:1909.12326 (2020). Yuang Jiang, Shiqiang Wang, Bong Jun Ko, Wei-Han Lee, and Leandros Tassiulas. 2020. Model Pruning Enables Efficient Federated Learning on Edge Devices. arXiv preprint arXiv:1909.12326 (2020)."},{"key":"e_1_3_2_2_9_1","volume-title":"Proceedings of the National Academy of Sciences","author":"Kirkpatrick James","year":"2017","unstructured":"James Kirkpatrick , Razvan Pascanu , Neil Rabinowitz , Joel Veness , Guillaume Desjardins , Andrei A Rusu , Kieran Milan , John Quan , Tiago Ramalho , Agnieszka Grabska-Barwinska , and et al. 2017. Overcoming catastrophic forgetting in neural nets . Proceedings of the National Academy of Sciences Mar 2017 , 114 (13) (2017). James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei ARusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, and et al. 2017. Overcoming catastrophic forgetting in neural nets. Proceedings of the National Academy of Sciences Mar 2017, 114 (13) (2017)."},{"key":"e_1_3_2_2_10_1","volume-title":"Federated optimization: Distributed machine learning for on-device intelligence. arXiv preprint arXiv:1610.02527","author":"Kone\u010dny Jakub","year":"2016","unstructured":"Jakub Kone\u010dny , H Brendan McMahan , Daniel Ramage , and Peter Richt\u00e1rik . 2016. Federated optimization: Distributed machine learning for on-device intelligence. arXiv preprint arXiv:1610.02527 ( 2016 ). Jakub Kone\u010dny, H Brendan McMahan, Daniel Ramage, and Peter Richt\u00e1rik. 2016. Federated optimization: Distributed machine learning for on-device intelligence. arXiv preprint arXiv:1610.02527 (2016)."},{"key":"e_1_3_2_2_11_1","volume-title":"Ananda Theertha Suresh, and Dave Bacon","author":"Kone\u010dny Jakub","year":"2016","unstructured":"Jakub Kone\u010dny , H. Brendan McMahan , Felix X. Yu , Peter Richt\u00e1rik , Ananda Theertha Suresh, and Dave Bacon . 2016 . Federated Learning : Strategies for Improving Communication Efficiency . arXiv preprint arXiv:1610.05492 (2016). Jakub Kone\u010dny, H. Brendan McMahan, Felix X. Yu, Peter Richt\u00e1rik, Ananda Theertha Suresh, and Dave Bacon. 2016. Federated Learning: Strategies for Improving Communication Efficiency. arXiv preprint arXiv:1610.05492 (2016)."},{"key":"e_1_3_2_2_12_1","unstructured":"Ang Li Jingwei Sun Binghui Wang Lin Duan Sicheng Li Yiran Chen and Hai Li. 2020. LotteryFL: Personalized and Communication-Efficient Federated Learning with Lottery Ticket Hypothesis on Non-IID Datasets. arXiv preprint arXiv:2008.03371 (2020).  Ang Li Jingwei Sun Binghui Wang Lin Duan Sicheng Li Yiran Chen and Hai Li. 2020. LotteryFL: Personalized and Communication-Efficient Federated Learning with Lottery Ticket Hypothesis on Non-IID Datasets. arXiv preprint arXiv:2008.03371 (2020)."},{"key":"e_1_3_2_2_13_1","volume-title":"Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith.","author":"Li Tian","year":"2019","unstructured":"Tian Li , Anit Kumar Sahu , Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith. 2019 . Federated Optimization in Heterogeneous Networks . arXiv preprint arXiv:1812.06127 (2019). Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith. 2019. Federated Optimization in Heterogeneous Networks. arXiv preprint arXiv:1812.06127 (2019)."},{"key":"e_1_3_2_2_14_1","volume-title":"Dynamic Model Pruning with Feedback. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=SJem8lSFwB","author":"Lin Tao","year":"2020","unstructured":"Tao Lin , Sebastian U. Stich , Luis Barba , Daniil Dmitriev , and Martin Jaggi . 2020 . Dynamic Model Pruning with Feedback. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=SJem8lSFwB Tao Lin, Sebastian U. Stich, Luis Barba, Daniil Dmitriev, and Martin Jaggi. 2020. Dynamic Model Pruning with Feedback. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=SJem8lSFwB"},{"key":"e_1_3_2_2_15_1","unstructured":"H Brendan McMahan Eider Moore Daniel Ramage Seth Hampson etal 2016. Communication-efficient learning of deep networks from decentralized data. arXiv preprint arXiv:1602.05629 (2016).  H Brendan McMahan Eider Moore Daniel Ramage Seth Hampson et al. 2016. Communication-efficient learning of deep networks from decentralized data. arXiv preprint arXiv:1602.05629 (2016)."},{"key":"e_1_3_2_2_16_1","volume-title":"Proceedings of the36thInternational Conference on MachineLearning","author":"Mohri Mehryar","year":"2019","unstructured":"Mehryar Mohri , Gary Sivek , and Ananda Theertha Suresh . 2019 . Agnostic Federated Learning . Proceedings of the36thInternational Conference on MachineLearning , Long Beach, California, PMLR 97 (2019). Mehryar Mohri, Gary Sivek, and Ananda Theertha Suresh. 2019. Agnostic Federated Learning. Proceedings of the36thInternational Conference on MachineLearning, Long Beach, California, PMLR 97 (2019)."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01158"},{"volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","key":"e_1_3_2_2_18_1","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , Luca Antiga , Alban Desmaison , Andreas Kopf , Edward Yang , Zachary DeVito , Martin Raison , Alykhan Tejani , Sasank Chilamkurthy , Benoit Steiner , Lu Fang , Junjie Bai , and Soumith Chintala . 2019. PyTorch: An Imperative Style , High-Performance Deep Learning Library . In Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.). Curran Associates, Inc., 8024-8035. http:\/\/papers.neurips.cc\/paper\/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.). Curran Associates, Inc., 8024-8035. http:\/\/papers.neurips.cc\/paper\/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf"},{"key":"e_1_3_2_2_19_1","first-page":"1","article-title":"Smartphone ownership and internet usage continues to climb in emerging economies","volume":"22","author":"Jacob Poushter","year":"2016","unstructured":"Jacob Poushter et al. 2016 . Smartphone ownership and internet usage continues to climb in emerging economies . Pew Research Center 22 (2016), 1 -- 44 . Jacob Poushter et al. 2016. Smartphone ownership and internet usage continues to climb in emerging economies. Pew Research Center 22 (2016), 1--44.","journal-title":"Pew Research Center"},{"key":"e_1_3_2_2_20_1","volume-title":"International Conference on Artificial Intelligence and Statistics. 2021--2031","author":"Reisizadeh Amirhossein","year":"2020","unstructured":"Amirhossein Reisizadeh , Aryan Mokhtari , Hamed Hassani , Ali Jadbabaie , and Ramtin Pedarsani . 2020 . Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization . In International Conference on Artificial Intelligence and Statistics. 2021--2031 . Amirhossein Reisizadeh, Aryan Mokhtari, Hamed Hassani, Ali Jadbabaie, and Ramtin Pedarsani. 2020. Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization. In International Conference on Artificial Intelligence and Statistics. 2021--2031."},{"key":"e_1_3_2_2_21_1","volume-title":"Progressive neural networks. arXiv preprint arXiv:1606.04671","author":"Rusu Andrei A","year":"2016","unstructured":"Andrei A Rusu , Neil C Rabinowitz , Guillaume Desjardins , Hubert Soyer , James Kirkpatrick , Koray Kavukcuoglu , Razvan Pascanu , and Raia Hadsell . 2016. Progressive neural networks. arXiv preprint arXiv:1606.04671 ( 2016 ). Andrei A Rusu, Neil C Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray Kavukcuoglu, Razvan Pascanu, and Raia Hadsell. 2016. Progressive neural networks. arXiv preprint arXiv:1606.04671 (2016)."},{"key":"e_1_3_2_2_22_1","volume-title":"Continual Learning with Deep Generative Replay. 31st Conference on Neural Information Processing Systems (NIPS 2017)","author":"Shin Hanul","year":"2017","unstructured":"Hanul Shin , Jung-Kwon Lee , Jaehong Kim , and Jiwon Kim . 2017 . Continual Learning with Deep Generative Replay. 31st Conference on Neural Information Processing Systems (NIPS 2017) (2017). Hanul Shin, Jung-Kwon Lee, Jaehong Kim, and Jiwon Kim. 2017. Continual Learning with Deep Generative Replay. 31st Conference on Neural Information Processing Systems (NIPS 2017) (2017)."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1177\/0047287514535847"},{"key":"e_1_3_2_2_24_1","volume-title":"Federated Learning with Matched Averaging. International Conference on Learning Representations","author":"Wang Hongyi","year":"2020","unstructured":"Hongyi Wang , Mikhail Yurochkin , Yuekai Sun , Dimitris Papailiopoulos , and Yasaman Khazaeni . 2020 . Federated Learning with Matched Averaging. International Conference on Learning Representations (2020). Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, and Yasaman Khazaeni. 2020. Federated Learning with Matched Averaging. International Conference on Learning Representations (2020)."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3298981"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.643"},{"key":"e_1_3_2_2_27_1","volume-title":"Lifelong learning with dynamically expandable networks. arXiv preprint arXiv:1708.01547","author":"Yoon Jaehong","year":"2017","unstructured":"Jaehong Yoon , Eunho Yang , Jeongtae Lee , and Sung Ju Hwang . 2017. Lifelong learning with dynamically expandable networks. arXiv preprint arXiv:1708.01547 ( 2017 ). Jaehong Yoon, Eunho Yang, Jeongtae Lee, and Sung Ju Hwang. 2017. Lifelong learning with dynamically expandable networks. arXiv preprint arXiv:1708.01547 (2017)."},{"key":"e_1_3_2_2_28_1","volume-title":"Incremental Network Quantization: Towards Lossless CNNs with Low-precision Weights. International Conference on Learning Representations","author":"Zhou Aojun","year":"2017","unstructured":"Aojun Zhou , Anbang Yao , Yiwen Guo , Lin Xu , and Yurong Chen . 2017 . Incremental Network Quantization: Towards Lossless CNNs with Low-precision Weights. International Conference on Learning Representations (2017). Aojun Zhou, Anbang Yao, Yiwen Guo, Lin Xu, and Yurong Chen. 2017. Incremental Network Quantization: Towards Lossless CNNs with Low-precision Weights. International Conference on Learning Representations (2017)."},{"key":"e_1_3_2_2_29_1","volume-title":"To prune, or not to prune: exploring the efficacy of pruning for model compression. arXiv preprint arXiv:1710.01878","author":"Zhu Michael","year":"2017","unstructured":"Michael Zhu and Suyog Gupta . 2017. To prune, or not to prune: exploring the efficacy of pruning for model compression. arXiv preprint arXiv:1710.01878 ( 2017 ). Michael Zhu and Suyog Gupta. 2017. To prune, or not to prune: exploring the efficacy of pruning for model compression. arXiv preprint arXiv:1710.01878 (2017)."}],"event":{"name":"CoNEXT '20: The 16th International Conference on emerging Networking EXperiments and Technologies","sponsor":["SIGCOMM ACM Special Interest Group on Data Communication"],"location":"Barcelona Spain","acronym":"CoNEXT '20"},"container-title":["Proceedings of the 1st Workshop on Distributed Machine Learning"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3426745.3431335","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3426745.3431335","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:01:45Z","timestamp":1750197705000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3426745.3431335"}},"subtitle":["Federated Weight Recovery"],"short-title":[],"issued":{"date-parts":[[2020,12]]},"references-count":29,"alternative-id":["10.1145\/3426745.3431335","10.1145\/3426745"],"URL":"https:\/\/doi.org\/10.1145\/3426745.3431335","relation":{},"subject":[],"published":{"date-parts":[[2020,12]]},"assertion":[{"value":"2020-12-01","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}