{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T04:44:17Z","timestamp":1742964257833,"version":"3.40.3"},"publisher-location":"Cham","reference-count":23,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030438227"},{"type":"electronic","value":"9783030438234"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-43823-4_25","type":"book-chapter","created":{"date-parts":[[2020,3,27]],"date-time":"2020-03-27T22:02:35Z","timestamp":1585346555000},"page":"293-300","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Distributed Learning of Neural Networks with One Round of Communication"],"prefix":"10.1007","author":[{"given":"Mike","family":"Izbicki","sequence":"first","affiliation":[]},{"given":"Christian R.","family":"Shelton","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,3,28]]},"reference":[{"key":"25_CR1","unstructured":"Han, J., Liu, Q.: Bootstrap model aggregation for distributed statistical learning. In: NeurIPS (2016)"},{"key":"25_CR2","unstructured":"Izbicki, M.: Divide and conquer algorithms for faster machine learning. PhD thesis, UC Riverside (2017)"},{"key":"25_CR3","doi-asserted-by":"crossref","unstructured":"Izbicki, M., Shelton, C.R.: Distributed learning of non-convex linear models with one round of communication. In: ECML-PKDD (2019)","DOI":"10.1007\/978-3-030-46147-8_12"},{"key":"25_CR4","unstructured":"Jaggi, M., et al.: Communication-efficient distributed dual coordinate ascent. In: NeurIPS (2014)"},{"key":"25_CR5","unstructured":"Jordan, M.I., Lee, J.D., Yang, Y.: Communication-efficient distributed statistical inference. arXiv preprint arXiv:1605.07689 (2016)"},{"key":"25_CR6","unstructured":"Kamp, M., Boley, M., Missura, O., G\u00e4rtner, T.: Effective parallelisation for machine learning. In: NeurIPS (2017)"},{"key":"25_CR7","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"393","DOI":"10.1007\/978-3-030-10925-7_24","volume-title":"Machine Learning and Knowledge Discovery in Databases","author":"M Kamp","year":"2019","unstructured":"Kamp, M., et al.: Efficient decentralized deep learning by dynamic model averaging. In: Berlingerio, M., Bonchi, F., G\u00e4rtner, T., Hurley, N., Ifrim, G. (eds.) ECML PKDD 2018. LNCS (LNAI), vol. 11051, pp. 393\u2013409. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-10925-7_24"},{"key":"25_CR8","unstructured":"Kone\u010dn\u1ef3, J., McMahan, H.B., Yu, F.X., Richt\u00e1rik, P., Suresh, A.T., Bacon, D.: Federated learning: strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492 (2016)"},{"key":"25_CR9","unstructured":"LeCun, Y., et al.: Lenet-5, convolutional neural networks"},{"issue":"5","key":"25_CR10","first-page":"115","volume":"18","author":"JD Lee","year":"2017","unstructured":"Lee, J.D., Liu, Q., Sun, Y., Taylor, J.E.: Communication-efficient sparse regression. JMLR 18(5), 115\u2013144 (2017)","journal-title":"JMLR"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"Li, M., Andersen, D.G., Park, J.W.: Scaling distributed machine learning with the parameter server. In: OSDI (2014)","DOI":"10.1145\/2640087.2644155"},{"key":"25_CR12","unstructured":"Liu, Q., Ihler, A.T.: Distributed estimation, information loss and exponential families. In: NeurIPS (2014)"},{"key":"25_CR13","unstructured":"Ma, C., Smith, V., Jaggi, M., Jordan, M.I., Richt\u00e1rik, P., Tak\u00e1\u010d, M.: Adding vs. averaging in distributed primal-dual optimization. In: ICML (2015)"},{"key":"25_CR14","unstructured":"McDonald, R., Mohri, M., Silberman, N., Walker, D., Mann, G.S.: Efficient large-scale distributed training of conditional maximum entropy models. In: NeurIPS (2009)"},{"key":"25_CR15","unstructured":"McMahan, H.B., Moore, E., Ramage, D., Arcas, B.A.y.: Federated learning of deep networks using model averaging. CoRR (2016)"},{"key":"25_CR16","unstructured":"McMahan, H.B., Moore, E., Ramage, D., Samson, S., Arcas, B.A.y.: Communication-efficient learning of deep networks from decentralized data (2017)"},{"issue":"4","key":"25_CR17","first-page":"379","volume":"5","author":"JD Rosenblatt","year":"2016","unstructured":"Rosenblatt, J.D., Nadler, B.: On the optimality of averaging in distributed statistical learning. Inf. Infer. 5(4), 379\u2013404 (2016)","journal-title":"Inf. Infer."},{"issue":"1","key":"25_CR18","first-page":"8590","volume":"18","author":"V Smith","year":"2018","unstructured":"Smith, V., Forte, S., Ma, C., Tak\u00e1\u010d, M., Jordan, M.I., Jaggi, M.: Cocoa: a general framework for communication-efficient distributed optimization. JMLR 18(1), 8590\u20138638 (2018)","journal-title":"JMLR"},{"key":"25_CR19","doi-asserted-by":"crossref","unstructured":"Wang, S.: A sharper generalization bound for divide-and-conquer ridge regression. In: AAAI (2019)","DOI":"10.1609\/aaai.v33i01.33015305"},{"key":"25_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Wainwright, M.J., Duchi, J.C.: Communication-efficient algorithms for statistical optimization. In: NeurIPS (2012)","DOI":"10.1109\/CDC.2012.6426691"},{"key":"25_CR21","unstructured":"Zhang, Y., Duchi, J.C., Wainwright,. M.J.: Divide and conquer kernel ridge regression. In: COLT (2013)"},{"key":"25_CR22","doi-asserted-by":"crossref","unstructured":"Zhao, S.-Y., Xiang, R., Shi, Y.-H., Gao, P., Li, W.-J.: Scalable composite optimization for learning on spark. In: AAAI, Scope (2017)","DOI":"10.1609\/aaai.v31i1.10920"},{"key":"25_CR23","unstructured":"Zinkevich, M., Weimer, M., Li, L., Smola, A.J.: Parallelized stochastic gradient descent. In: NeurIPS (2010)"}],"container-title":["Communications in Computer and Information Science","Machine Learning and Knowledge Discovery in Databases"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-43823-4_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,13]],"date-time":"2024-02-13T01:15:38Z","timestamp":1707786938000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-43823-4_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030438227","9783030438234"],"references-count":23,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-43823-4_25","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"28 March 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECML PKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"W\u00fcrzburg","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 September 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 September 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecml2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/ecmlpkdd2019.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"733","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"130","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"18% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.04","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ECML PKDD Workshops Information: single-blind review, submissions: 200, full papers accepted: 70, short papers accepted: 46","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}