{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T16:16:23Z","timestamp":1762272983067,"version":"3.37.3"},"reference-count":26,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,11,7]],"date-time":"2022-11-07T00:00:00Z","timestamp":1667779200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,11,7]],"date-time":"2022-11-07T00:00:00Z","timestamp":1667779200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"General Program of National Natural Science Foundation of China","award":["61871388"],"award-info":[{"award-number":["61871388"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2023,4]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Federated learning (FL) is a new distributed learning framework that is different from traditional distributed machine learning: (1) differences in communication, computing, and storage performance among devices (device heterogeneity), (2) differences in data distribution and data volume (data heterogeneity), and (3) high communication consumption. Under heterogeneous conditions, the data distribution of clients varies greatly, which leads to the problem that the convergence speed of the training model decreases and the training model cannot converge to the global optimal solution. In this work, an FL algorithm based on stratified sampling and regularization (FedSSAR) is proposed. In FedSSAR, a density-based clustering method is used to divide the overall client into different clusters, then, some available clients are proportionally extracted from different clusters to participate in training which realizes unbiased sampling for the overall client and reduces the aggregation weight variance of the client. At the same time, when calculating the model local loss function, we limit the update direction of the model by a regular term, so that heterogeneous clients are optimized in the globally optimal direction. We prove the convergence of FedSSAR theoretically and experimentally, and demonstrate the superiority of FedSSAR by comparing it with other FL algorithms on public datasets.<\/jats:p>","DOI":"10.1007\/s40747-022-00895-3","type":"journal-article","created":{"date-parts":[[2022,11,7]],"date-time":"2022-11-07T05:02:47Z","timestamp":1667797367000},"page":"2081-2099","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Federated learning based on stratified sampling and regularization"],"prefix":"10.1007","volume":"9","author":[{"given":"Chenyang","family":"Lu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4606-6171","authenticated-orcid":false,"given":"Wubin","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Rui","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Su","family":"Deng","sequence":"additional","affiliation":[]},{"given":"Yahui","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,7]]},"reference":[{"issue":"3","key":"895_CR1","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/MSP.2020.2975749","volume":"37","author":"T Li","year":"2020","unstructured":"Li T, Sahu AK, Talwalkar A, Smith V (2020) Federated learning: challenges, methods, and future directions. IEEE Signal Process Mag 37(3):50\u201360","journal-title":"IEEE Signal Process Mag"},{"key":"895_CR2","unstructured":"Mcmahan HB, Moore E, Ramage D, Hampson S, Arcas B (2016) Communication-efficient learning of deep networks from decentralized data. arXiv:1602.05629"},{"key":"895_CR3","unstructured":"Mcmahan HB, Ramage D, Talwar K, Zhang L (2017) Learning differentially private recurrent language models. arXiv:1710.06963"},{"issue":"2","key":"895_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3298981","volume":"10","author":"Q Yang","year":"2019","unstructured":"Yang Q, Liu Y, Chen T, Tong Y (2019) Federated machine learning. ACM Trans Intell Syst Technol 10(2):1\u201319","journal-title":"ACM Trans Intell Syst Technol"},{"key":"895_CR5","unstructured":"Hsieh K, Phanishayee A, Mutlu O, Gibbons PB (2019) The non-IID data quagmire of decentralized machine learning. arXiv:1910.00189"},{"key":"895_CR6","unstructured":"Li T, Sahu AK, Zaheer M, Sanjabi M, Talwalkar A, Smith V (2018) Federated optimization in heterogeneous networks. arXiv:1812.06127"},{"key":"895_CR7","unstructured":"Hard A, Rao K, Mathews R, Beaufays F, Ramage D (2018) Federated learning for mobile keyboard prediction. arXiv:1811.03604"},{"key":"895_CR8","unstructured":"Yang T, Andrew G, Eichner H, Sun H, Li W, Kong N, Ramage D, Franoise B (2018) Applied federated learning: improving google keyboard query suggestions. arXiv:1812.02903"},{"key":"895_CR9","unstructured":"Karimireddy SP, Kale S, Mohri M et al (2020) Scaffold: stochastic controlled averaging for federated learning. In: International Conference on  Machine Learning. PMLR, pp 5132\u20135143"},{"issue":"1","key":"895_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1561\/2200000016","volume":"3","author":"S Boyd","year":"2010","unstructured":"Boyd S, Parikh N, Hu EC, Peleato B, EcKstein J (2010) Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Found Trends Machine Learn 3(1):1\u2013122","journal-title":"Found Trends Machine Learn"},{"issue":"1","key":"895_CR11","first-page":"165","volume":"13","author":"O Dekel","year":"2012","unstructured":"Dekel O, Gilad-Bachrach R, Shamir O, Xiao L (2012) Optimal distributed online prediction using mini-batches. J Machine Learn Res 13(1):165\u2013202","journal-title":"J Machine Learn Res"},{"issue":"1","key":"895_CR12","first-page":"2657","volume":"17","author":"P Richt\u00e1rik","year":"2016","unstructured":"Richt\u00e1rik P, Tak\u00e1\u010d M (2016) Distributed coordinate descent method for learning with big data. J Machine Learn Res 17(1):2657\u20132681","journal-title":"J Machine Learn Res"},{"key":"895_CR13","volume-title":"Deep learning with Elastic Averaging SGD","author":"S Zhang","year":"2014","unstructured":"Zhang S, Choromanska A, Lecun Y (2014) Deep learning with Elastic Averaging SGD. MIT Press, Cambridge"},{"key":"895_CR14","doi-asserted-by":"crossref","unstructured":"Bonawitz K, Ivanov V, Kreuter B et al (2017) Practical secure aggregation for privacy-preserving machine learning. In: Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, pp 1175\u20131191","DOI":"10.1145\/3133956.3133982"},{"key":"895_CR15","unstructured":"Bonawitz K et al (2021) Towards federated learning at scale: system design. arXiv:1912.04977"},{"key":"895_CR16","unstructured":"Mohri M, Sivek G, Suresh AT (2019) Agnostic federated learning. arXiv:1902.01046"},{"key":"895_CR17","doi-asserted-by":"crossref","unstructured":"Hu H, Wang D, Wu C (2020) Distributed machine learning through heterogeneous edge systems. In: Proceedings of the AAAI Conference on Artificial Intelligence, Vol 34, no. 05, pp 7179\u20137186","DOI":"10.1609\/aaai.v34i05.6207"},{"key":"895_CR18","doi-asserted-by":"crossref","unstructured":"Peter Kairouz H, McMahan B et al (2021) Advances and open problems in federated learning. arXiv:1912.04977","DOI":"10.1561\/9781680837896"},{"key":"895_CR19","unstructured":"Zhao Y, Li M, Lai L, Suda N, Civin D, Chandra V (2018) Federated learning with non-IID data. arXiv:1806.00582"},{"key":"895_CR20","unstructured":"Yan Y, Niu C, Ding Y, Zheng Z, Wu F, Chen G, Tang S, Wu Z (2020) Distributed non-convex optimization with sublinear speedup under intermittent client availability. arXiv:2002.07399"},{"key":"895_CR21","unstructured":"Acar DAE, Zhao Y, Navarro RM et al (2021) Federated learning based on dynamic regularization. arXiv:2111.04263"},{"key":"895_CR22","doi-asserted-by":"crossref","unstructured":"Huang Y, Chu L, Zhou Z et al (2020) Personalized cross-silo federated learning on non-IID data. arXiv:2007.03797","DOI":"10.1609\/aaai.v35i9.16960"},{"key":"895_CR23","unstructured":"Ghosh A, Chung J, Dong Y, Ramchandran K (2020) An efficient framework for clustered federated learning. arXiv:2006.04088"},{"issue":"8","key":"895_CR24","doi-asserted-by":"publisher","first-page":"3710","DOI":"10.1109\/TNNLS.2020.3015958","volume":"32","author":"F Sattler","year":"2021","unstructured":"Sattler F, Muller KR, Samek W (2021) Clustered federated learning: model-agnostic distributed multitask optimization under privacy constraints. IEEE Trans Neural Netw Learn Syst 32(8):3710\u20133722","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"895_CR25","unstructured":"Fraboni Y, Vidal R, Kameni L et al (2021) Clustered sampling: low-variance and improved representativity for clients selection in federated learning. In: International Conference on Machine Learning. PMLR, pp 3407\u20133416"},{"key":"895_CR26","first-page":"49","volume":"28","author":"MB Ankerst","year":"1999","unstructured":"Ankerst MB, Kriegel MM, Sander HPJ (1999) Optics: ordering points to identify the clustering structure. SIGMOD Record Spec Interest Group Manage Data 28:49\u201360","journal-title":"SIGMOD Record Spec Interest Group Manage Data"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00895-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-022-00895-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00895-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,18]],"date-time":"2023-04-18T09:45:33Z","timestamp":1681811133000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-022-00895-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,11,7]]},"references-count":26,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,4]]}},"alternative-id":["895"],"URL":"https:\/\/doi.org\/10.1007\/s40747-022-00895-3","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"type":"print","value":"2199-4536"},{"type":"electronic","value":"2198-6053"}],"subject":[],"published":{"date-parts":[[2022,11,7]]},"assertion":[{"value":"1 July 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 October 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 November 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}