{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T11:40:19Z","timestamp":1755776419144,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":84,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,3,30]],"date-time":"2025-03-30T00:00:00Z","timestamp":1743292800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,3,30]]},"DOI":"10.1145\/3689031.3717467","type":"proceedings-article","created":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T06:25:20Z","timestamp":1742970320000},"page":"1317-1333","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Hourglass: Enabling Efficient Split Federated Learning with Data Parallelism"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2607-4556","authenticated-orcid":false,"given":"Qiang","family":"He","sequence":"first","affiliation":[{"name":"Huazhong University of Science and Technology, Wuhan, China, Swinburne University of Technology, Melbourne, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5967-5692","authenticated-orcid":false,"given":"Kaibin","family":"Wang","sequence":"additional","affiliation":[{"name":"Swinburne University of Technology, Melbourne, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8496-7224","authenticated-orcid":false,"given":"Zeqian","family":"Dong","sequence":"additional","affiliation":[{"name":"Swinburne University of Technology, Melbourne, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6766-7913","authenticated-orcid":false,"given":"Liang","family":"Yuan","sequence":"additional","affiliation":[{"name":"University of Adelaide, Adelaide, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5455-3792","authenticated-orcid":false,"given":"Feifei","family":"Chen","sequence":"additional","affiliation":[{"name":"Deakin University, Melbourne, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3934-7605","authenticated-orcid":false,"given":"Hai","family":"Jin","sequence":"additional","affiliation":[{"name":"Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7868-5471","authenticated-orcid":false,"given":"Yun","family":"Yang","sequence":"additional","affiliation":[{"name":"Swinburne University of Technology, Melbourne, Australia"}]}],"member":"320","published-online":{"date-parts":[[2025,3,30]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Post training 4-bit quantization of convolutional networks for rapid-deployment. Advances in Neural Information Processing Systems 32","author":"Banner Ron","year":"2019","unstructured":"Ron Banner, Yury Nahshan, and Daniel Soudry. 2019. Post training 4-bit quantization of convolutional networks for rapid-deployment. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_1_2_1","volume-title":"19th USENIX Symposium on Networked Systems Design and Implementation (NSDI 22)","author":"Bhardwaj Romil","year":"2022","unstructured":"Romil Bhardwaj, Zhengxu Xia, Ganesh Ananthanarayanan, Junchen Jiang, Yuanchao Shu, Nikolaos Karianakis, Kevin Hsieh, Paramvir Bahl, and Ion Stoica. 2022. Ekya: Continuous learning of video analytics models on edge compute servers. In 19th USENIX Symposium on Networked Systems Design and Implementation (NSDI 22). 119--135."},{"key":"e_1_3_2_1_3_1","volume-title":"Proceedings of machine learning and systems 1","author":"Bonawitz Keith","year":"2019","unstructured":"Keith Bonawitz, Hubert Eichner, Wolfgang Grieskamp, Dzmitry Huba, Alex Ingerman, Vladimir Ivanov, Chloe Kiddon, Jakub Kone\u010dny, Stefano Mazzocchi, Brendan McMahan, et al. 2019. Towards federated learning at scale: System design. Proceedings of machine learning and systems 1 (2019), 374--388."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3570361.3592505"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00309"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458817.3476211"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550289"},{"key":"e_1_3_2_1_9_1","first-page":"6476","article-title":"Smyrf-efficient attention using asymmetric clustering","volume":"33","author":"Daras Giannis","year":"2020","unstructured":"Giannis Daras, Nikita Kitaev, Augustus Odena, and Alexandros G Dimakis. 2020. Smyrf-efficient attention using asymmetric clustering. Advances in Neural Information Processing Systems 33 (2020), 6476--6489.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_10_1","volume-title":"Cinic-10 is not imagenet or cifar-10. arXiv preprint arXiv:1810.03505","author":"Darlow Luke N","year":"2018","unstructured":"Luke N Darlow, Elliot J Crowley, Antreas Antoniou, and Amos J Storkey. 2018. Cinic-10 is not imagenet or cifar-10. arXiv preprint arXiv:1810.03505 (2018)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/2020408.2020578"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/997817.997857"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2020.2976475"},{"key":"e_1_3_2_1_14_1","unstructured":"Hugging Face. 2024. Quantization. https:\/\/huggingface.co\/docs\/optimum\/en\/concept_guides\/quantization"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01544"},{"key":"e_1_3_2_1_16_1","volume-title":"29th USENIX Security Symposium (USENIX Security 20)","author":"Fang Minghong","year":"2020","unstructured":"Minghong Fang, Xiaoyu Cao, Jinyuan Jia, and Neil Gong. 2020. Local model poisoning attacks to {byzantine-robust} federated learning. In 29th USENIX Security Symposium (USENIX Security 20). 1605--1622."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00197"},{"key":"e_1_3_2_1_18_1","first-page":"19586","article-title":"An efficient framework for clustered federated learning","volume":"33","author":"Ghosh Avishek","year":"2020","unstructured":"Avishek Ghosh, Jichan Chung, Dong Yin, and Kannan Ramchandran. 2020. An efficient framework for clustered federated learning. Advances in Neural Information Processing Systems 33 (2020), 19586--19597.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_19_1","volume-title":"ICML 2021 workshop on federated learning for user privacy and data confidentiality. ICML Board.","author":"Han Dong-Jun","year":"2021","unstructured":"Dong-Jun Han, Hasnain Irshad Bhatti, Jungmoon Lee, and Jaekyun Moon. 2021. Accelerating federated learning with split learning on locally generated losses. In ICML 2021 workshop on federated learning for user privacy and data confidentiality. ICML Board."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM53939.2023.10229027"},{"key":"e_1_3_2_1_21_1","volume-title":"Convergence Analysis of Split Federated Learning on Heterogeneous Data. arXiv preprint arXiv:2402.15166","author":"Han Pengchao","year":"2024","unstructured":"Pengchao Han, Chao Huang, Geng Tian, Ming Tang, and Xin Liu. 2024. Convergence Analysis of Split Federated Learning on Heterogeneous Data. arXiv preprint arXiv:2402.15166 (2024)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00852"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01197"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3373376.3378530"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00291"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243"},{"key":"e_1_3_2_1_28_1","first-page":"814","article-title":"Papaya: Practical, private, and scalable federated learning","volume":"4","author":"Huba Dzmitry","year":"2022","unstructured":"Dzmitry Huba, John Nguyen, Kshitiz Malik, Ruiyu Zhu, Mike Rabbat, Ashkan Yousefpour, Carole-Jean Wu, Hongyuan Zhan, Pavel Ustinov, Harish Srinivas, et al. 2022. Papaya: Practical, private, and scalable federated learning. Proceedings of Machine Learning and Systems 4 (2022), 814--832.","journal-title":"Proceedings of Machine Learning and Systems"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00286"},{"key":"e_1_3_2_1_30_1","volume-title":"Vision Xformers: Efficient attention for image classification. arXiv preprint arXiv:2107.02239","author":"Jeevan Pranav","year":"2021","unstructured":"Pranav Jeevan and Amit Sethi. 2021. Vision Xformers: Efficient attention for image classification. arXiv preprint arXiv:2107.02239 (2021)."},{"key":"e_1_3_2_1_31_1","volume-title":"International Conference on Machine Learning. PMLR, 5201--5212","author":"Kairouz Peter","year":"2021","unstructured":"Peter Kairouz, Ziyu Liu, and Thomas Steinke. 2021. The distributed discrete gaussian mechanism for federated learning with secure aggregation. In International Conference on Machine Learning. PMLR, 5201--5212."},{"key":"e_1_3_2_1_32_1","volume-title":"International Conference on Machine Learning. PMLR, 5132--5143","author":"Karimireddy Sai Praneeth","year":"2020","unstructured":"Sai Praneeth Karimireddy, Satyen Kale, Mehryar Mohri, Sashank Reddi, Sebastian Stich, and Ananda Theertha Suresh. 2020. Scaffold: Stochastic controlled averaging for federated learning. In International Conference on Machine Learning. PMLR, 5132--5143."},{"key":"e_1_3_2_1_33_1","volume-title":"2019 International Conference on Machine Learning. PMLR, 3509--3518","author":"Koratana Animesh","year":"2019","unstructured":"Animesh Koratana, Daniel Kang, Peter Bailis, and Matei Zaharia. 2019. Lit: Learned intermediate representation training for model compression. In 2019 International Conference on Machine Learning. PMLR, 3509--3518."},{"key":"e_1_3_2_1_34_1","unstructured":"Alex Krizhevsky and Geoffrey Hinton. 2009. CIFAR-10 dataset. https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE.2019.00027"},{"key":"e_1_3_2_1_36_1","volume-title":"Oort: Efficient federated learning via guided participant selection. In 15th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 21). 19--35.","author":"Lai Fan","year":"2021","unstructured":"Fan Lai, Xiangfeng Zhu, Harsha V Madhyastha, and Mosharaf Chowdhury. 2021. Oort: Efficient federated learning via guided participant selection. In 15th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 21). 19--35."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3495243.3517017"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00995"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE53745.2022.00077"},{"key":"e_1_3_2_1_40_1","volume-title":"International Conference on Machine Learning. PMLR","author":"Li Songze","year":"2023","unstructured":"Songze Li, Duanyi Yao, and Jin Liu. 2023. Fedvs: Straggler-resilient and privacy-preserving vertical federated learning for split models. In International Conference on Machine Learning. PMLR, 20296--20311."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2020.2975749"},{"key":"e_1_3_2_1_42_1","volume-title":"Proceedings of Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V. Sze (Eds.)","volume":"2","author":"Li Tian","year":"2020","unstructured":"Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith. 2020. Federated optimization in heterogeneous networks. In Proceedings of Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V. Sze (Eds.), Vol. 2. 429--450."},{"key":"e_1_3_2_1_43_1","volume-title":"On the convergence of fedavg on non-iid data. arXiv preprint arXiv:1907.02189","author":"Li Xiang","year":"2019","unstructured":"Xiang Li, Kaixuan Huang, Wenhao Yang, Shusen Wang, and Zhihua Zhang. 2019. On the convergence of fedavg on non-iid data. arXiv preprint arXiv:1907.02189 (2019)."},{"key":"e_1_3_2_1_44_1","volume-title":"Convergence analysis of sequential federated learning on heterogeneous data. Advances in Neural Information Processing Systems 36","author":"Li Yipeng","year":"2024","unstructured":"Yipeng Li and Xinchen Lyu. 2024. Convergence analysis of sequential federated learning on heterogeneous data. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE60146.2024.00164"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3620678.3624651"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICC40277.2020.9148862"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2019.2946538"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM42981.2021.9488679"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2858232"},{"key":"e_1_3_2_1_51_1","first-page":"5972","article-title":"No fear of heterogeneity: Classifier calibration for federated learning with non-iid data","volume":"34","author":"Luo Mi","year":"2021","unstructured":"Mi Luo, Fei Chen, Dapeng Hu, Yifan Zhang, Jian Liang, and Jiashi Feng. 2021. No fear of heterogeneity: Classifier calibration for federated learning with non-iid data. Advances in Neural Information Processing Systems 34 (2021), 5972--5984.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_52_1","first-page":"43097","article-title":"Structured federated learning through clustered additive modeling","volume":"36","author":"Ma Jie","year":"2023","unstructured":"Jie Ma, Tianyi Zhou, Guodong Long, Jing Jiang, and Chengqi Zhang. 2023. Structured federated learning through clustered additive modeling. Advances in Neural Information Processing Systems 36 (2023), 43097--43107.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_53_1","unstructured":"Brendan McMahan Eider Moore Daniel Ramage Seth Hampson and Blaise Aguera y Arcas. 2017. Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics. PMLR 1273--1282."},{"key":"e_1_3_2_1_54_1","volume-title":"International Conference on Machine Learning. PMLR, 7197--7206","author":"Nagel Markus","year":"2020","unstructured":"Markus Nagel, Rana Ali Amjad, Mart Van Baalen, Christos Louizos, and Tijmen Blankevoort. 2020. Up or down? Adaptive rounding for post-training quantization. In International Conference on Machine Learning. PMLR, 7197--7206."},{"key":"e_1_3_2_1_55_1","unstructured":"Maxime Oquab Timoth\u00e9e Darcet Th\u00e9o Moutakanni Huy Vo Marc Szafraniec Vasil Khalidov Pierre Fernandez Daniel Haziza Francisco Massa Alaaeldin El-Nouby et al. 2023. Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193 (2023)."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458864.3467681"},{"key":"e_1_3_2_1_57_1","volume-title":"20th USENIX Symposium on Networked Systems Design and Implementation. 973--994","author":"Padmanabhan Arthi","year":"2023","unstructured":"Arthi Padmanabhan, Neil Agarwal, Anand Iyer, Ganesh Ananthanarayanan, Yuanchao Shu, Nikolaos Karianakis, Guoqing Harry Xu, and Ravi Netravali. 2023. Gemel: Model merging for memory-efficient, real-time video analytics at the edge. In 20th USENIX Symposium on Networked Systems Design and Implementation. 973--994."},{"key":"e_1_3_2_1_58_1","unstructured":"Paperswithcode. 2012. Image classification on CIFAR-10. https:\/\/paperswithcode.com\/sota\/image-classification-on-cifar-10"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3460120.3485259"},{"key":"e_1_3_2_1_60_1","volume-title":"International Conference on Learning Representations.","author":"Polino Antonio","year":"2018","unstructured":"Antonio Polino, Razvan Pascanu, and Dan Alistarh. 2018. Model compression via distillation and quantization. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_61_1","volume-title":"Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints","author":"Sattler Felix","year":"2020","unstructured":"Felix Sattler, Klaus-Robert M\u00fcller, and Wojciech Samek. 2020. Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints. IEEE transactions on neural networks and learning systems 32, 8 (2020), 3710--3722."},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3603269.3604830"},{"key":"e_1_3_2_1_63_1","volume-title":"Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)."},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2020.3000372"},{"key":"e_1_3_2_1_65_1","volume-title":"International conference on machine learning. PMLR, 10096--10106","author":"Tan Mingxing","year":"2021","unstructured":"Mingxing Tan and Quoc Le. 2021. Efficientnetv2: Smaller models and faster training. In International conference on machine learning. PMLR, 10096--10106."},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE51399.2021.00224"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i8.20825"},{"key":"e_1_3_2_1_68_1","volume-title":"Mlp-mixer: An all-mlp architecture for vision. Advances in neural information processing systems 34","author":"Tolstikhin Ilya O","year":"2021","unstructured":"Ilya O Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Andreas Steiner, Daniel Keysers, Jakob Uszkoreit, et al. 2021. Mlp-mixer: An all-mlp architecture for vision. Advances in neural information processing systems 34 (2021), 24261--24272."},{"key":"e_1_3_2_1_69_1","volume-title":"International Conference on Machine Learning. PMLR, 10347--10357","author":"Touvron Hugo","year":"2021","unstructured":"Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herv\u00e9 J\u00e9gou. 2021. Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning. PMLR, 10347--10357."},{"key":"e_1_3_2_1_70_1","volume-title":"Attention is all you need. Advances in neural information processing systems 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_71_1","volume-title":"Fluid: Mitigating stragglers in federated learning using invariant dropout. Advances in Neural Information Processing Systems","author":"Wang Irene","year":"2024","unstructured":"Irene Wang, Prashant Nair, and Divya Mahajan. 2024. Fluid: Mitigating stragglers in federated learning using invariant dropout. Advances in Neural Information Processing Systems (2024), 13--36."},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583347"},{"key":"e_1_3_2_1_73_1","volume-title":"Proceedings of the ACM Web Conference","author":"Wang Kaibin","year":"2023","unstructured":"Kaibin Wang, Qiang He, Feifei Chen, Hai Jin, and Yun Yang. 2023. Fed-Edge: Accelerating Edge-Assisted Federated Learning. In Proceedings of the ACM Web Conference 2023. 2895--2904."},{"key":"e_1_3_2_1_74_1","volume-title":"Training deep neural networks with 8-bit floating point numbers. Advances in Neural Information Processing Systems 31","author":"Wang Naigang","year":"2018","unstructured":"Naigang Wang, Jungwook Choi, Daniel Brand, Chia-Yu Chen, and Kailash Gopalakrishnan. 2018. Training deep neural networks with 8-bit floating point numbers. Advances in Neural Information Processing Systems 31 (2018)."},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM42981.2021.9488756"},{"key":"e_1_3_2_1_76_1","volume-title":"Speech commands: A dataset for limited-vocabulary speech recognition. arXiv preprint arXiv:1804.03209","author":"Warden Pete","year":"2018","unstructured":"Pete Warden. 2018. Speech commands: A dataset for limited-vocabulary speech recognition. arXiv preprint arXiv:1804.03209 (2018)."},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2020.2988575"},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2023.3242704"},{"key":"e_1_3_2_1_79_1","volume-title":"2024 USENIX Annual Technical Conference (ATC 24)","author":"Xu Mengwei","year":"2024","unstructured":"Mengwei Xu, Dongqi Cai, Yaozong Wu, Xiang Li, and Shangguang Wang. 2024. FwdLLM: Efficient federated finetuning of large language models with perturbed inferences. In 2024 USENIX Annual Technical Conference (ATC 24). 579--596."},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE.2017.145"},{"key":"e_1_3_2_1_81_1","volume-title":"Character-level convolutional networks for text classification. Advances in neural information processing systems 28","author":"Zhang Xiang","year":"2015","unstructured":"Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. Advances in neural information processing systems 28 (2015)."},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539231"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA56546.2023.10071077"},{"key":"e_1_3_2_1_84_1","volume-title":"Discrimination-aware channel pruning for deep neural networks. Advances in Neural Information Processing Systems","author":"Zhuang Zhuangwei","year":"2018","unstructured":"Zhuangwei Zhuang, Mingkui Tan, Bohan Zhuang, Jing Liu, Yong Guo, Qingyao Wu, Junzhou Huang, and Jinhui Zhu. 2018. Discrimination-aware channel pruning for deep neural networks. Advances in Neural Information Processing Systems (2018), 31--43."}],"event":{"name":"EuroSys '25: Twentieth European Conference on Computer Systems","sponsor":["SIGOPS ACM Special Interest Group on Operating Systems"],"location":"Rotterdam Netherlands","acronym":"EuroSys '25"},"container-title":["Proceedings of the Twentieth European Conference on Computer Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3689031.3717467","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3689031.3717467","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T11:21:29Z","timestamp":1755775289000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3689031.3717467"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,30]]},"references-count":84,"alternative-id":["10.1145\/3689031.3717467","10.1145\/3689031"],"URL":"https:\/\/doi.org\/10.1145\/3689031.3717467","relation":{},"subject":[],"published":{"date-parts":[[2025,3,30]]},"assertion":[{"value":"2025-03-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}