{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T08:09:45Z","timestamp":1743149385488,"version":"3.40.3"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031599323"},{"type":"electronic","value":"9783031599330"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-59933-0_9","type":"book-chapter","created":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T11:01:49Z","timestamp":1716807709000},"page":"118-137","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Data Summarization for\u00a0Federated Learning"],"prefix":"10.1007","author":[{"given":"Julianna","family":"Devillers","sequence":"first","affiliation":[]},{"given":"Olivier","family":"Brun","sequence":"additional","affiliation":[]},{"given":"Balakrishna J.","family":"Prabhu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,25]]},"reference":[{"key":"9_CR1","unstructured":"Flower github repository (2023). https:\/\/github.com\/adap\/flower. Accessed 30 Aug 2023"},{"key":"9_CR2","unstructured":"Beutel, D.J., et al.: Flower: a friendly federated learning research framework. arXiv preprint arXiv:2007.14390 (2020)"},{"key":"9_CR3","doi-asserted-by":"crossref","unstructured":"Cui, L., Su, X., Zhou, Y., Zhang, L.: ClusterGrad: adaptive gradient compression by clustering in federated learning. In: 2020 IEEE Global Communications Conference 2020, pp. 1\u20137 (2020)","DOI":"10.1109\/GLOBECOM42002.2020.9322527"},{"key":"9_CR4","unstructured":"Devillers, J.: Data summarization methods for energy efficient federated learning. Master\u2019s thesis internship report, ISAE-SUPAERO (2023). https:\/\/github.com\/juliannadvl\/FL-DS\/blob\/main\/rapport_stage.pdf"},{"issue":"4","key":"9_CR5","doi-asserted-by":"publisher","first-page":"634","DOI":"10.1145\/285055.285059","volume":"45","author":"U Feige","year":"1998","unstructured":"Feige, U.: A threshold of ln n for approximating set cover. J. ACM 45(4), 634\u2013652 (1998). https:\/\/doi.org\/10.1145\/285055.285059","journal-title":"J. ACM"},{"key":"9_CR6","unstructured":"Haddadpour, F., Kamani, M.M., Mokhtari, A., Mahdavi, M.: Federated learning with compression: unified analysis and sharp guarantees. In: PMLR, vol. 130, pp. 2350\u20132358 (2021)"},{"key":"9_CR7","unstructured":"Jiang, P., Agrawal, G.: A linear speedup analysis of distributed deep learning with sparse and quantized communication. In: NeurIPS (2018)"},{"key":"9_CR8","unstructured":"Karimireddy, S.P., Kale, S., Mohri, M., Reddi, S., Stich, S., Suresh, A.T.: Scaffold: stochastic controlled averaging for federated learning. In: III, H.D., Singh, A. (eds.) Proceedings of the 37th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol.\u00a0119, pp. 5132\u20135143. PMLR (2020). https:\/\/proceedings.mlr.press\/v119\/karimireddy20a.html"},{"key":"9_CR9","unstructured":"Khaled, A., Mishchenko, K., Richt\u00e1rik, P.: Better communication complexity for local SGD. CoRR abs\/1909.04746 (2019). http:\/\/arxiv.org\/abs\/1909.04746"},{"key":"9_CR10","unstructured":"Krause, A., Guestrin, C.: Near-optimal nonmyopic value of information in graphical models. In: Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence, UAI 2005, pp. 324\u2013331. AUAI Press, Arlington (2005)"},{"key":"9_CR11","doi-asserted-by":"publisher","unstructured":"Li, T., Sahu, A.K., Zaheer, M., Sanjabi, M., Talwalkar, A., Smith, V.: Federated optimization in heterogeneous networks (2018). https:\/\/doi.org\/10.48550\/ARXIV.1812.06127. https:\/\/arxiv.org\/abs\/1812.06127","DOI":"10.48550\/ARXIV.1812.06127"},{"key":"9_CR12","unstructured":"Li, X., Huang, K., Yang, W., Wang, S., Zhang, Z.: On the convergence of FedAvg on Non-IID data (2020)"},{"key":"9_CR13","unstructured":"McMahan, B., Moore, E., Ramage, D., Hampson, S., Aguera y Arcas, B.: Communication-efficient learning of deep networks from decentralized data. In: Singh, A., Zhu, J. (eds.) Proceedings of the 20th International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol.\u00a054, pp. 1273\u20131282. PMLR (2017). https:\/\/proceedings.mlr.press\/v54\/mcmahan17a.html"},{"key":"9_CR14","unstructured":"Mirzasoleiman, B.: Big data summarization using submodular functions. Ph.D. thesis, ETH Zurich (2017)"},{"key":"9_CR15","unstructured":"Mirzasoleiman, B., Bilmes, J., Leskovec, J.: Coresets for data-efficient training of machine learning models. In: Proceedings of the 37th International Conference on Machine Learning, ICML 2020. JMLR.org (2020)"},{"key":"9_CR16","doi-asserted-by":"publisher","first-page":"265","DOI":"10.1007\/BF01588971","volume":"14","author":"GL Nemhauser","year":"1978","unstructured":"Nemhauser, G.L., Wolsey, L.A., Fisher, M.L.: An analysis of approximations for maximizing submodular set functions - I. Math. Program. 14, 265\u2013294 (1978)","journal-title":"Math. Program."},{"key":"9_CR17","unstructured":"Patterson, D., et al.: Carbon emissions and large neural network training (2021)"},{"key":"9_CR18","unstructured":"Qiu, X., et al.: A first look into the carbon footprint of federated learning (2023)"},{"key":"9_CR19","unstructured":"Reddi, S.J., et al.: Adaptive federated optimization. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=LkFG3lB13U5"},{"key":"9_CR20","doi-asserted-by":"crossref","unstructured":"Schwartz, R., Dodge, J., Smith, N.A., Etzioni, O.: Green AI (2019)","DOI":"10.1145\/3381831"},{"key":"9_CR21","doi-asserted-by":"publisher","unstructured":"Strubell, E., Ganesh, A., McCallum, A.: Energy and policy considerations for deep learning in NLP. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 3645\u20133650. Association for Computational Linguistics, Florence (2019). https:\/\/doi.org\/10.18653\/v1\/P19-1355. https:\/\/aclanthology.org\/P19-1355","DOI":"10.18653\/v1\/P19-1355"},{"key":"9_CR22","unstructured":"Weinberg, M.: Lecture notes in advanced algorithm design - lecture 7: submodular functions, lov\u00e1sz extension and minimization (2022). https:\/\/www.cs.princeton.edu\/~hy2\/teaching\/fall22-cos521\/notes\/SFM.pdf. Accessed 19 Apr 2023"},{"key":"9_CR23","doi-asserted-by":"crossref","unstructured":"Yu, H., Yang, S., Zhu, S.: Parallel restarted SGD with faster convergence and less communication: demystifying why model averaging works for deep learning (2018)","DOI":"10.1609\/aaai.v33i01.33015693"},{"key":"9_CR24","unstructured":"Yuan, X.T., Li, P.: On convergence of FedProx: local dissimilarity invariant bounds, non-smoothness and beyond (2022)"}],"container-title":["Lecture Notes in Computer Science","Machine Learning for Networking"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-59933-0_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T11:03:21Z","timestamp":1716807801000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-59933-0_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031599323","9783031599330"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-59933-0_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"25 May 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MLN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Machine Learning for Networking","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Paris","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mln2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.adda-association.org\/mln-2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"16","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}