{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T16:41:13Z","timestamp":1773247273085,"version":"3.50.1"},"reference-count":52,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"7","license":[{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Hong Kong Research Grants Council","award":["AoE\/E-601\/22-R"],"award-info":[{"award-number":["AoE\/E-601\/22-R"]}]},{"name":"NSFC\/RGC Collaborative Research Scheme","award":["CRS_HKUST603\/22"],"award-info":[{"award-number":["CRS_HKUST603\/22"]}]},{"DOI":"10.13039\/501100004377","name":"Hong Kong Polytechnic University","doi-asserted-by":"publisher","award":["P0038174"],"award-info":[{"award-number":["P0038174"]}],"id":[{"id":"10.13039\/501100004377","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. on Mobile Comput."],"published-print":{"date-parts":[[2024,7]]},"DOI":"10.1109\/tmc.2023.3338021","type":"journal-article","created":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T19:27:53Z","timestamp":1701372473000},"page":"7572-7584","source":"Crossref","is-referenced-by-count":20,"title":["MimiC: Combating Client Dropouts in Federated Learning by Mimicking Central Updates"],"prefix":"10.1109","volume":"23","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7881-4723","authenticated-orcid":false,"given":"Yuchang","family":"Sun","sequence":"first","affiliation":[{"name":"Department of Electronic and Computer Engineering, The Hong Kong University of Science and Technology, Clear Water Bay, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5646-8679","authenticated-orcid":false,"given":"Yuyi","family":"Mao","sequence":"additional","affiliation":[{"name":"Department of Electrical and Electronic Engineering, Hong Kong Polytechnic University, Hung Hom, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5222-1898","authenticated-orcid":false,"given":"Jun","family":"Zhang","sequence":"additional","affiliation":[{"name":"Department of Electronic and Computer Engineering, The Hong Kong University of Science and Technology, Clear Water Bay, Hong Kong"}]}],"member":"263","reference":[{"key":"ref1","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","volume-title":"Proc. Int. Conf. Artif. Intell. Statist.","author":"McMahan"},{"key":"ref2","first-page":"374","article-title":"Towards federated learning at scale: System design","volume-title":"Proc. Mach. Learn. Syst.","author":"Bonawitz"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.cie.2020.106854"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2021.3080078"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/GLOBECOM46510.2021.9685654"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2023.3253853"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1561\/9781680837896"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2020.2986024"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2019.2918951"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/MCOM.001.1900103"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/IPDPS47924.2020.00031"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3517207.3526969"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2021.3095077"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1561\/9781680837056"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/JSAIT.2020.2991361"},{"key":"ref16","article-title":"Straggler-resilient distributed machine learning with dynamic backup workers","author":"Xiong","year":"2021"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2022.3224590"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2021.3106104"},{"key":"ref19","article-title":"Achieving linear speedup with partial worker participation in non-IID federated learning","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Yang"},{"key":"ref20","article-title":"On the convergence of FedAvg on non-IID data","author":"Li","year":"2019"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-28996-5_4"},{"key":"ref22","article-title":"Distributed non-convex optimization with sublinear speedup under intermittent client availability","author":"Yan","year":"2020"},{"key":"ref23","first-page":"12052","article-title":"Fast federated learning in the presence of arbitrary device unavailability","volume-title":"Proc. 35th Conf. Adv. Neural Inf. Process. Syst.","author":"Gu"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2020.3031503"},{"key":"ref25","first-page":"3403","article-title":"Towards flexible device participation in federated learning","volume-title":"Proc Int. Conf. Artif. Intell. Statist.","author":"Ruan"},{"key":"ref26","first-page":"5677","article-title":"A unified analysis of federated learning with arbitrary client participation","volume-title":"Proc. 35th Conf. Adv. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2022.3195073"},{"issue":"72","key":"ref28","first-page":"1","article-title":"Redundancy techniques for straggler mitigation in distributed optimization and learning","volume":"20","author":"Karakus","year":"2019","journal-title":"J. Mach. Learn. Res."},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2020.3036961"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/twc.2023.3334732"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TCOMM.2023.3244243"},{"key":"ref32","first-page":"10533","article-title":"DRes-FL: Dropout-resilient secure federated learning for non-IID clients via secret data sharing","volume-title":"Proc. 31st Conf. Adv. Neural Inf. Process. Syst.","author":"Shao"},{"key":"ref33","article-title":"Combating client dropout in federated learning via friend model substitution","author":"Wang","year":"2022"},{"key":"ref34","article-title":"Taming client dropout for distributed differential privacy in federated learning","author":"Jiang"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2022.3163592"},{"key":"ref36","first-page":"7611","article-title":"Tackling the objective inconsistency problem in heterogeneous federated optimization","volume-title":"Proc. 34th Conf. Adv. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref37","first-page":"429","article-title":"Federated optimization in heterogeneous networks","volume-title":"Proc. Mach. Learn. Syst","author":"Li"},{"key":"ref38","first-page":"5132","article-title":"SCAFFOLD: Stochastic controlled averaging for federated learning","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Karimireddy"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM48880.2022.9796818"},{"key":"ref40","article-title":"Local SGD converges fast and communicates little","author":"Stich","year":"2018"},{"key":"ref41","first-page":"11080","article-title":"Local SGD with periodic averaging: Tighter analysis and adaptive synchronization","volume-title":"Proc. 33rd Conf. Adv. Neural Inf. Process. Syst.","author":"Haddadpour"},{"key":"ref42","first-page":"4519","article-title":"Tighter theory for local SGD on identical and heterogeneous data","volume-title":"Proc. Int. Conf. Artif. Intell. Statist.","author":"Khaled"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1137\/16M1080173"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.5555\/2968826.2969010"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/TCOMM.2022.3151126"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3133982"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2021.3075439"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS54860.2022.00061"},{"key":"ref49","article-title":"Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms","author":"Xiao","year":"2017"},{"key":"ref50","article-title":"Learning multiple layers of features from tiny images","author":"Krizhevsky","year":"2009"},{"key":"ref51","article-title":"Federated learning on non-IID data silos: An experimental study","author":"Li","year":"2021"},{"key":"ref52","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2014"}],"container-title":["IEEE Transactions on Mobile Computing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7755\/10550110\/10336542.pdf?arnumber=10336542","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,25]],"date-time":"2024-06-25T21:25:34Z","timestamp":1719350734000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10336542\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7]]},"references-count":52,"journal-issue":{"issue":"7"},"URL":"https:\/\/doi.org\/10.1109\/tmc.2023.3338021","relation":{},"ISSN":["1536-1233","1558-0660","2161-9875"],"issn-type":[{"value":"1536-1233","type":"print"},{"value":"1558-0660","type":"electronic"},{"value":"2161-9875","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7]]}}}