{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T09:26:28Z","timestamp":1750325188668,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":24,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,6,24]],"date-time":"2021-06-24T00:00:00Z","timestamp":1624492800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"the Fundamental Research Funds for the Central Universities, and National Natural Science Foundation of China under grant numbers","award":["62032003, 61922017, and 61921003."],"award-info":[{"award-number":["62032003, 61922017, and 61921003."]}]},{"name":"National Key R&D Program of China","award":["2020YFB1805500"],"award-info":[{"award-number":["2020YFB1805500"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,6,25]]},"DOI":"10.1145\/3469116.3470009","type":"proceedings-article","created":{"date-parts":[[2021,6,24]],"date-time":"2021-06-24T10:10:05Z","timestamp":1624529405000},"page":"31-36","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Towards Ubiquitous Learning"],"prefix":"10.1145","author":[{"given":"Dongqi","family":"Cai","sequence":"first","affiliation":[{"name":"State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications, China"}]},{"given":"Qipeng","family":"Wang","sequence":"additional","affiliation":[{"name":"Key Lab of High Confidence Software Technologies (Peking University), MoE, China"}]},{"given":"Yuanqiang","family":"Liu","sequence":"additional","affiliation":[{"name":"Key Lab of High Confidence Software Technologies (Peking University), MoE, China"}]},{"given":"Yunxin","family":"Liu","sequence":"additional","affiliation":[{"name":"Institute for AI Industry Research (AIR), Tsinghua University, China"}]},{"given":"Shangguang","family":"Wang","sequence":"additional","affiliation":[{"name":"Shenzhen Research Institute, Beijing University of Posts and Telecommunications, China"}]},{"given":"Mengwei","family":"Xu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications, China"}]}],"member":"320","published-online":{"date-parts":[[2021,6,24]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Collaborative machine learning without centralized training data. https:\/\/ai.googleblog.com\/2017\/04\/federated-learning-collaborative.html","author":"Federated","year":"2017","unstructured":"Federated learning : Collaborative machine learning without centralized training data. https:\/\/ai.googleblog.com\/2017\/04\/federated-learning-collaborative.html , 2017 . Federated learning: Collaborative machine learning without centralized training data. https:\/\/ai.googleblog.com\/2017\/04\/federated-learning-collaborative.html, 2017."},{"key":"e_1_3_2_1_2_1","volume-title":"https:\/\/www.technologyreview.com\/2019\/12\/11\/131629\/apple-ai-personalizes-siri-federated-learning\/","author":"How","year":"2019","unstructured":"How apple personalizes siri without hoovering up your data. https:\/\/www.technologyreview.com\/2019\/12\/11\/131629\/apple-ai-personalizes-siri-federated-learning\/ , 2019 . How apple personalizes siri without hoovering up your data. https:\/\/www.technologyreview.com\/2019\/12\/11\/131629\/apple-ai-personalizes-siri-federated-learning\/, 2019."},{"key":"e_1_3_2_1_3_1","volume-title":"https:\/\/deeplearning4j.org\/","author":"Deep","year":"2021","unstructured":"Deep learning for java. https:\/\/deeplearning4j.org\/ , 2021 . Deep learning for java. https:\/\/deeplearning4j.org\/, 2021."},{"key":"e_1_3_2_1_4_1","volume-title":"Towards federated learning at scale: System design. arXiv preprint arXiv:1902.01046","author":"Bonawitz Keith","year":"2019","unstructured":"Keith Bonawitz , Hubert Eichner , Wolfgang Grieskamp , Dzmitry Huba , Alex Ingerman , Vladimir Ivanov , Chloe Kiddon , Jakub Kone\u010dny , Stefano Mazzocchi , H Brendan McMahan , Towards federated learning at scale: System design. arXiv preprint arXiv:1902.01046 , 2019 . Keith Bonawitz, Hubert Eichner, Wolfgang Grieskamp, Dzmitry Huba, Alex Ingerman, Vladimir Ivanov, Chloe Kiddon, Jakub Kone\u010dny, Stefano Mazzocchi, H Brendan McMahan, et al. Towards federated learning at scale: System design. arXiv preprint arXiv:1902.01046, 2019."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.5555\/3291168.3291211"},{"key":"e_1_3_2_1_6_1","volume-title":"Training deep nets with sublinear memory cost. arXiv preprint arXiv:1604.06174","author":"Chen Tianqi","year":"2016","unstructured":"Tianqi Chen , Bing Xu , Chiyuan Zhang , and Carlos Guestrin . Training deep nets with sublinear memory cost. arXiv preprint arXiv:1604.06174 , 2016 . Tianqi Chen, Bing Xu, Chiyuan Zhang, and Carlos Guestrin. Training deep nets with sublinear memory cost. arXiv preprint arXiv:1604.06174, 2016."},{"key":"e_1_3_2_1_7_1","volume-title":"End-to-end evaluation of federated learning and split learning for internet of things. arXiv preprint arXiv:2003.13376","author":"Gao Yansong","year":"2020","unstructured":"Yansong Gao , Minki Kim , Sharif Abuadbba , Yeonjae Kim , Chandra Thapa , Kyuyeon Kim , Seyit A Camtepe , Hyoungshick Kim , and Surya Nepal . End-to-end evaluation of federated learning and split learning for internet of things. arXiv preprint arXiv:2003.13376 , 2020 . Yansong Gao, Minki Kim, Sharif Abuadbba, Yeonjae Kim, Chandra Thapa, Kyuyeon Kim, Seyit A Camtepe, Hyoungshick Kim, and Surya Nepal. End-to-end evaluation of federated learning and split learning for internet of things. arXiv preprint arXiv:2003.13376, 2020."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.5555\/3323234.3323274"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3373376.3378530"},{"key":"e_1_3_2_1_10_1","volume-title":"et al. Mnn: A universal and efficient inference engine. arXiv preprint arXiv:2002.12418","author":"Jiang Xiaotang","year":"2020","unstructured":"Xiaotang Jiang , Huan Wang , Yiliu Chen , Ziqi Wu , Lichuan Wang , Bin Zou , Yafeng Yang , Zongyang Cui , Yu Cai , Tianhang Yu , et al. Mnn: A universal and efficient inference engine. arXiv preprint arXiv:2002.12418 , 2020 . Xiaotang Jiang, Huan Wang, Yiliu Chen, Ziqi Wu, Lichuan Wang, Bin Zou, Yafeng Yang, Zongyang Cui, Yu Cai, Tianhang Yu, et al. Mnn: A universal and efficient inference engine. arXiv preprint arXiv:2002.12418, 2020."},{"key":"e_1_3_2_1_11_1","volume-title":"Keith Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al. Advances and open problems in federated learning. arXiv preprint arXiv","author":"Kairouz Peter","year":"1912","unstructured":"Peter Kairouz , H Brendan McMahan , Brendan Avent , Aur\u00e9lien Bellet , Mehdi Bennis , Arjun Nitin Bhagoji , Keith Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al. Advances and open problems in federated learning. arXiv preprint arXiv 1912 .04977, 2019. Peter Kairouz, H Brendan McMahan, Brendan Avent, Aur\u00e9lien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Keith Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al. Advances and open problems in federated learning. arXiv preprint arXiv 1912.04977, 2019."},{"key":"e_1_3_2_1_12_1","first-page":"1273","volume-title":"Artificial Intelligence and Statistics","author":"McMahan Brendan","year":"2017","unstructured":"Brendan McMahan , Eider Moore , Daniel Ramage , Seth Hampson , and Blaise Aguera y Arcas . Communication-efficient learning of deep networks from decentralized data . In Artificial Intelligence and Statistics , pages 1273 -- 1282 . PMLR, 2017 . Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. Communication-efficient learning of deep networks from decentralized data. In Artificial Intelligence and Statistics, pages 1273--1282. PMLR, 2017."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3372224.3419188"},{"key":"e_1_3_2_1_14_1","volume-title":"Don't decay the learning rate, increase the batch size. arXiv preprint arXiv 1711.00489","author":"Smith Samuel L","year":"2017","unstructured":"Samuel L Smith , Pieter-Jan Kindermans , Chris Ying , and Quoc V Le . Don't decay the learning rate, increase the batch size. arXiv preprint arXiv 1711.00489 , 2017 . Samuel L Smith, Pieter-Jan Kindermans, Chris Ying, and Quoc V Le. Don't decay the learning rate, increase the batch size. arXiv preprint arXiv 1711.00489, 2017."},{"key":"e_1_3_2_1_15_1","volume-title":"Split learning for health: Distributed deep learning without sharing raw patient data. arXiv preprint arXiv","author":"Vepakomma Praneeth","year":"1812","unstructured":"Praneeth Vepakomma , Otkrist Gupta , Tristan Swedish , and Ramesh Raskar . Split learning for health: Distributed deep learning without sharing raw patient data. arXiv preprint arXiv 1812 .00564, 2018. Praneeth Vepakomma, Otkrist Gupta, Tristan Swedish, and Ramesh Raskar. Split learning for health: Distributed deep learning without sharing raw patient data. arXiv preprint arXiv 1812.00564, 2018."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447993.3448625"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3308558.3313591"},{"key":"e_1_3_2_1_18_1","volume-title":"A case for camera-as-a-service","author":"Xu Mengwei","year":"2021","unstructured":"Mengwei Xu , Yunxin Liu , and Xuanzhe Liu . A case for camera-as-a-service . IEEE Pervasive Computing , 2021 . Mengwei Xu, Yunxin Liu, and Xuanzhe Liu. A case for camera-as-a-service. IEEE Pervasive Computing, 2021."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287075"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2019.2893250"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386901.3388948"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3241539.3241563"},{"key":"e_1_3_2_1_23_1","volume-title":"Heterogeneity-aware federated learning. arXiv preprint arXiv.2006.06983","author":"Yang Chengxu","year":"2020","unstructured":"Chengxu Yang , QiPeng Wang , Mengwei Xu , Shangguang Wang , Kaigui Bian , and Xuanzhe Liu . Heterogeneity-aware federated learning. arXiv preprint arXiv.2006.06983 , 2020 . Chengxu Yang, QiPeng Wang, Mengwei Xu, Shangguang Wang, Kaigui Bian, and Xuanzhe Liu. Heterogeneity-aware federated learning. arXiv preprint arXiv.2006.06983, 2020."},{"key":"e_1_3_2_1_24_1","first-page":"515","volume-title":"Yuqi Wang, Yifan Xiong, et al. Hived: Sharing a {GPU} cluster for deep learning with guarantees. In 14th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 20)","author":"Zhao Hanyu","year":"2020","unstructured":"Hanyu Zhao , Zhenhua Han , Zhi Yang , Quanlu Zhang , Fan Yang , Lidong Zhou , Mao Yang , Francis CM Lau , Yuqi Wang, Yifan Xiong, et al. Hived: Sharing a {GPU} cluster for deep learning with guarantees. In 14th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 20) , pages 515 -- 532 , 2020 . Hanyu Zhao, Zhenhua Han, Zhi Yang, Quanlu Zhang, Fan Yang, Lidong Zhou, Mao Yang, Francis CM Lau, Yuqi Wang, Yifan Xiong, et al. Hived: Sharing a {GPU} cluster for deep learning with guarantees. In 14th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 20), pages 515--532, 2020."}],"event":{"name":"MobiSys '21: The 19th Annual International Conference on Mobile Systems, Applications, and Services","sponsor":["SIGMOBILE ACM Special Interest Group on Mobility of Systems, Users, Data and Computing","SIGOPS ACM Special Interest Group on Operating Systems"],"location":"Virtual WI USA","acronym":"MobiSys '21"},"container-title":["Proceedings of the 5th International Workshop on Embedded and Mobile Deep Learning"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3469116.3470009","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3469116.3470009","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:29Z","timestamp":1750191509000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3469116.3470009"}},"subtitle":["A First Measurement of On-Device Training Performance"],"short-title":[],"issued":{"date-parts":[[2021,6,24]]},"references-count":24,"alternative-id":["10.1145\/3469116.3470009","10.1145\/3469116"],"URL":"https:\/\/doi.org\/10.1145\/3469116.3470009","relation":{},"subject":[],"published":{"date-parts":[[2021,6,24]]},"assertion":[{"value":"2021-06-24","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}