{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T18:42:12Z","timestamp":1772822532285,"version":"3.50.1"},"reference-count":55,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/Crown.html"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Netw. Sci. Eng."],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/tnse.2025.3577910","type":"journal-article","created":{"date-parts":[[2025,6,6]],"date-time":"2025-06-06T13:43:13Z","timestamp":1749217393000},"page":"1-16","source":"Crossref","is-referenced-by-count":3,"title":["Straggler-resilient Federated Learning: Tackling Computation Heterogeneity with Layer-wise Partial Model Training in Mobile Edge Network"],"prefix":"10.1109","author":[{"given":"Hongda","family":"Wu","sequence":"first","affiliation":[{"name":"Department of Electrical Engineering and Computer Science, Lassonde School of Engineering, York University, Toronto, ON, Canada"}]},{"given":"Ping","family":"Wang","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Science, Lassonde School of Engineering, York University, Toronto, ON, Canada"}]},{"given":"C V Aswartha","family":"Narayana","sequence":"additional","affiliation":[{"name":"Department of Electrical Engineering and Computer Science, Lassonde School of Engineering, York University, Toronto, ON, Canada"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCC57788.2023.10233400"},{"key":"ref2","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","volume-title":"Proc. Artif. Intell. Statist. Conf.","author":"McMahan","year":"2017"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2019.2941458"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2020.3041388"},{"key":"ref5","first-page":"3252","article-title":"Error feedback fixes signSGD and other gradient compression schemes","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Karimireddy","year":"2019"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2019.2961673"},{"key":"ref7","first-page":"5132","article-title":"Scaffold: Stochastic controlled averaging for federated learning","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Karimireddy","year":"2020"},{"key":"ref8","first-page":"10351","article-title":"Towards understanding biased client selection in federated learning","volume-title":"Proc. Int. Conf. Artif. Intell. Statist.","author":"Cho","year":"2022"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TNSE.2022.3146399"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM48880.2022.9796935"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TCCN.2021.3084406"},{"key":"ref12","article-title":"Personalized federated learning with first order model optimization","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Zhang","year":"2021"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM48880.2022.9796733"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207469"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/tit.2022.3192506"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICC.2019.8761315"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM42981.2021.9488679"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2020.3035770"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2021.3052681"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2024789118"},{"key":"ref21","article-title":"On the opportunities and risks of foundation models","author":"Bommasani","year":"2021"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"issue":"7","key":"ref24","article-title":"Distilling the knowledge in a neural network","volume":"2","author":"Hinton","year":"2015"},{"key":"ref25","first-page":"2351","article-title":"Ensemble distillation for robust model fusion in federated learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Lin","year":"2020"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2022.3231527"},{"key":"ref27","first-page":"14068","article-title":"Group knowledge transfer: Federated learning of large cnns at the edge","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"He","year":"2020"},{"key":"ref28","first-page":"12878","article-title":"Data-free knowledge distillation for heterogeneous federated learning","volume-title":"Proc. 38th Int. Conf. Mach. Learn.","author":"Zhu","year":"2021"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE53745.2022.00062"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3166101"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/LWC.2022.3149783"},{"key":"ref32","first-page":"2021","article-title":"HeteroFL: Computation and communication efficient federated learning for heterogeneous clients","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Diao"},{"key":"ref33","first-page":"12876","article-title":"FjORD: Fair and accurate federated learning under heterogeneous targets with ordered dropout","volume-title":"Proc. 35th Int. Conf. Neural Inf. Process. Syst.","author":"Horvath","year":"2021"},{"key":"ref34","first-page":"29677","article-title":"Fedrolex: Model-heterogeneous federated learning with rolling sub-model extraction","volume-title":"Proc. 6th Int. Conf. Neural Inf. Process. Syst.","author":"Alam","year":"2022"},{"key":"ref35","article-title":"Slimmable neural networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Yu","year":"2019"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00189"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.14778\/3529337.3529343"},{"key":"ref38","first-page":"2021","article-title":"No fear of heterogeneity: Classifier calibration for federated learning with non-IID data","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Luo"},{"key":"ref39","article-title":"Rapid learning or feature reuse? towards understanding the effectiveness of MAML","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Raghu","year":"2020"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2024.3353377"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671897"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/tccn.2025.3553315"},{"key":"ref43","first-page":"3299","article-title":"meProp: Sparsified back propagation for accelerated deep learning with reduced overfitting","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Sun","year":"2017"},{"key":"ref44","article-title":"Split learning for health: Distributed deep learning without sharing raw patient data","author":"Vepakomma","year":"2018"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i8.20825"},{"key":"ref46","first-page":"1047","article-title":"Split learning for distributed collaborative training of deep learning models in health informatics","volume-title":"Proc. AMIA Annu. Symp.","author":"Li","year":"2024"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2023.3309633"},{"key":"ref48","article-title":"On the convergence of FEDAVG on non-IID data","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Li","year":"2020"},{"key":"ref49","article-title":"On the convergence of local descent methods in federated learning","author":"Haddadpour","year":"2019"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/IPDPS47924.2020.00033"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2023.3297790"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2019.2904348"},{"key":"ref53","article-title":"Local SGD converges fast and communicates little","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Stich","year":"2019"},{"key":"ref54","first-page":"429","article-title":"Federated optimization in heterogeneous networks","volume-title":"Proc. Mach. Learn. Syst.","author":"Li","year":"2020"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299173"}],"container-title":["IEEE Transactions on Network Science and Engineering"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6488902\/6930788\/11027799.pdf?arnumber=11027799","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T04:40:15Z","timestamp":1761367215000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11027799\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":55,"URL":"https:\/\/doi.org\/10.1109\/tnse.2025.3577910","relation":{},"ISSN":["2327-4697","2334-329X"],"issn-type":[{"value":"2327-4697","type":"electronic"},{"value":"2334-329X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}