{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T17:30:17Z","timestamp":1771954217330,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":72,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,23]],"date-time":"2023-10-23T00:00:00Z","timestamp":1698019200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,23]]},"DOI":"10.1145\/3600006.3613139","type":"proceedings-article","created":{"date-parts":[[2023,10,3]],"date-time":"2023-10-03T14:44:17Z","timestamp":1696344257000},"page":"331-347","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":18,"title":["PIT: Optimization of Dynamic Sparse Deep Learning Models via Permutation Invariant Transformation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-6449-8972","authenticated-orcid":false,"given":"Ningxin","family":"Zheng","sequence":"first","affiliation":[{"name":"Microsoft Research, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1327-4882","authenticated-orcid":false,"given":"Huiqiang","family":"Jiang","sequence":"additional","affiliation":[{"name":"Microsoft Research, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0557-1104","authenticated-orcid":false,"given":"Quanlu","family":"Zhang","sequence":"additional","affiliation":[{"name":"Microsoft Research, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2880-7100","authenticated-orcid":false,"given":"Zhenhua","family":"Han","sequence":"additional","affiliation":[{"name":"Microsoft Research, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-9524-5476","authenticated-orcid":false,"given":"Lingxiao","family":"Ma","sequence":"additional","affiliation":[{"name":"Microsoft Research, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3518-5212","authenticated-orcid":false,"given":"Yuqing","family":"Yang","sequence":"additional","affiliation":[{"name":"Microsoft Research, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0378-060X","authenticated-orcid":false,"given":"Fan","family":"Yang","sequence":"additional","affiliation":[{"name":"Microsoft Research, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4939-0436","authenticated-orcid":false,"given":"Chengruidong","family":"Zhang","sequence":"additional","affiliation":[{"name":"Microsoft Research, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-8131-7439","authenticated-orcid":false,"given":"Lili","family":"Qiu","sequence":"additional","affiliation":[{"name":"Microsoft Research, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-6455-3898","authenticated-orcid":false,"given":"Mao","family":"Yang","sequence":"additional","affiliation":[{"name":"Microsoft Research, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7258-3116","authenticated-orcid":false,"given":"Lidong","family":"Zhou","sequence":"additional","affiliation":[{"name":"Microsoft Research, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2023,10,23]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"https:\/\/github.com\/allenai\/longformer","year":"2020","unstructured":"Longformer. https:\/\/github.com\/allenai\/longformer , 2020 . Longformer. https:\/\/github.com\/allenai\/longformer, 2020."},{"key":"e_1_3_2_1_2_1","volume-title":"https:\/\/developer.nvidia.com\/blog\/accelerating-inference-with-sparsity-using-ampere-and-tensorrt\/","author":"Accelerating","year":"2021","unstructured":"Accelerating inference with sparsity using the nvidia ampere architecture and nvidia tensorrt. https:\/\/developer.nvidia.com\/blog\/accelerating-inference-with-sparsity-using-ampere-and-tensorrt\/ , 2021 . Accelerating inference with sparsity using the nvidia ampere architecture and nvidia tensorrt. https:\/\/developer.nvidia.com\/blog\/accelerating-inference-with-sparsity-using-ampere-and-tensorrt\/, 2021."},{"key":"e_1_3_2_1_3_1","volume-title":"the cuda sparse matrix library. https:\/\/docs.nvidia.com\/cuda\/cusparse\/index.html","author":"The","year":"2021","unstructured":"The api reference guide for cusparse , the cuda sparse matrix library. https:\/\/docs.nvidia.com\/cuda\/cusparse\/index.html , 2021 . The api reference guide for cusparse, the cuda sparse matrix library. https:\/\/docs.nvidia.com\/cuda\/cusparse\/index.html, 2021."},{"key":"e_1_3_2_1_4_1","volume-title":"https:\/\/docs.nvidia.com\/cuda\/cuda-c-programming-guide\/index.html#wmma","author":"Cuda","year":"2021","unstructured":"Cuda c++ programming guide. https:\/\/docs.nvidia.com\/cuda\/cuda-c-programming-guide\/index.html#wmma , 2021 . Cuda c++ programming guide. https:\/\/docs.nvidia.com\/cuda\/cuda-c-programming-guide\/index.html#wmma, 2021."},{"key":"e_1_3_2_1_5_1","volume-title":"A high-performance cuda library for sparse matrix-matrix multiplication. https:\/\/docs.nvidia.com\/cuda\/cusparselt\/index.html","year":"2021","unstructured":"cusparselt : A high-performance cuda library for sparse matrix-matrix multiplication. https:\/\/docs.nvidia.com\/cuda\/cusparselt\/index.html , 2021 . cusparselt: A high-performance cuda library for sparse matrix-matrix multiplication. https:\/\/docs.nvidia.com\/cuda\/cusparselt\/index.html, 2021."},{"key":"e_1_3_2_1_6_1","volume-title":"https:\/\/github.com\/openai\/blocksparse.git","author":"Openai","year":"2021","unstructured":"Openai block sparse. https:\/\/github.com\/openai\/blocksparse.git , 2021 . Openai block sparse. https:\/\/github.com\/openai\/blocksparse.git, 2021."},{"key":"e_1_3_2_1_7_1","volume-title":"https:\/\/github.com\/tensor-compiler\/taco\/tree\/oopsla2020","author":"Reproducing","year":"2021","unstructured":"Reproducing oopsla 2020 results. https:\/\/github.com\/tensor-compiler\/taco\/tree\/oopsla2020 , 2021 . Reproducing oopsla 2020 results. https:\/\/github.com\/tensor-compiler\/taco\/tree\/oopsla2020, 2021."},{"key":"e_1_3_2_1_8_1","volume-title":"https:\/\/github.com\/ROCmSoftwar_ePlatform\/hipSPARSE","author":"Rocm","year":"2021","unstructured":"Rocm sparse marshalling library. https:\/\/github.com\/ROCmSoftwar_ePlatform\/hipSPARSE , 2021 . Rocm sparse marshalling library. https:\/\/github.com\/ROCmSoftwar_ePlatform\/hipSPARSE, 2021."},{"key":"e_1_3_2_1_9_1","volume-title":"https:\/\/github.com\/openai\/triton.git","year":"2021","unstructured":"Triton. https:\/\/github.com\/openai\/triton.git , 2021 . Triton. https:\/\/github.com\/openai\/triton.git, 2021."},{"key":"e_1_3_2_1_10_1","volume-title":"https:\/\/github.com\/apache\/tvm\/blob\/254563a3140cf63fe77a46058688209de3aa213c\/python\/tvm\/topi\/cuda\/sparse.py#L96","author":"Tvm","year":"2021","unstructured":"Tvm sparsity code. https:\/\/github.com\/apache\/tvm\/blob\/254563a3140cf63fe77a46058688209de3aa213c\/python\/tvm\/topi\/cuda\/sparse.py#L96 , 2021 . Tvm sparsity code. https:\/\/github.com\/apache\/tvm\/blob\/254563a3140cf63fe77a46058688209de3aa213c\/python\/tvm\/topi\/cuda\/sparse.py#L96, 2021."},{"key":"e_1_3_2_1_11_1","volume-title":"https:\/\/github.com\/microsoft\/DeepSpeed","year":"2022","unstructured":"Deepspeed. https:\/\/github.com\/microsoft\/DeepSpeed , 2022 . Deepspeed. https:\/\/github.com\/microsoft\/DeepSpeed, 2022."},{"key":"e_1_3_2_1_12_1","unstructured":"Riyadh Baghdadi Abdelkader Nadir Debbagh Kamel Abdous Fatima Zohra Benhamida Alex Renda Jonathan Elliott Frankle Michael Carbin and Saman Amarasinghe. Tiramisu: A polyhedral compiler for dense and sparse deep learning. ArXiv preprint abs\/2005.04091 2020.  Riyadh Baghdadi Abdelkader Nadir Debbagh Kamel Abdous Fatima Zohra Benhamida Alex Renda Jonathan Elliott Frankle Michael Carbin and Saman Amarasinghe. Tiramisu: A polyhedral compiler for dense and sparse deep learning. ArXiv preprint abs\/2005.04091 2020."},{"key":"e_1_3_2_1_13_1","volume-title":"Citeseer","author":"Bell Nathan","year":"2008","unstructured":"Nathan Bell and Michael Garland . Efficient sparse matrix-vector multiplication on cuda. Technical report , Citeseer , 2008 . Nathan Bell and Michael Garland. Efficient sparse matrix-vector multiplication on cuda. Technical report, Citeseer, 2008."},{"key":"e_1_3_2_1_14_1","volume-title":"Longformer: The long-document transformer. ArXiv preprint, abs\/2004.05150","author":"Beltagy Iz","year":"2020","unstructured":"Iz Beltagy , Matthew E Peters , and Arman Cohan . Longformer: The long-document transformer. ArXiv preprint, abs\/2004.05150 , 2020 . Iz Beltagy, Matthew E Peters, and Arman Cohan. Longformer: The long-document transformer. ArXiv preprint, abs\/2004.05150, 2020."},{"key":"e_1_3_2_1_15_1","first-page":"2023","article-title":"Efficient gpu kernels for n:m-sparse weights in deep learning","author":"Bin Lin","year":"2023","unstructured":"Lin Bin , Zheng Ningxin , Wang Yang , Cao Shijie , Ma Lingxiao , Zhang Quanlu , Zhu Yi , Cao Ting , Xue Jilong , Yang Yuqing , and Yang Fan . Efficient gpu kernels for n:m-sparse weights in deep learning . MLSys 2023 , 2023 . Lin Bin, Zheng Ningxin, Wang Yang, Cao Shijie, Ma Lingxiao, Zhang Quanlu, Zhu Yi, Cao Ting, Xue Jilong, Yang Yuqing, and Yang Fan. Efficient gpu kernels for n:m-sparse weights in deep learning. MLSys2023, 2023.","journal-title":"MLSys"},{"key":"e_1_3_2_1_16_1","volume-title":"Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020","author":"Brown Tom B.","year":"2020","unstructured":"Tom B. Brown , Benjamin Mann , Nick Ryder , Melanie Subbiah , Jared Kaplan , Prafulla Dhariwal , Arvind Neelakantan , Pranav Shyam , Girish Sastry , Amanda Askell , Sandhini Agarwal , Ariel Herbert-Voss , Gretchen Krueger , Tom Henighan , Rewon Child , Aditya Ramesh , Daniel M. Ziegler , Jeffrey Wu , Clemens Winter , Christopher Hesse , Mark Chen , Eric Sigler , Mateusz Litwin , Scott Gray , Benjamin Chess , Jack Clark , Christopher Berner , Sam McCandlish , Alec Radford , Ilya Sutskever , and Dario Amodei . Language models are few-shot learners. In Hugo Larochelle, Marc'Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors , Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020 , NeurIPS 2020 , December 6--12, 2020, virtual, 2020. Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language models are few-shot learners. In Hugo Larochelle, Marc'Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6--12, 2020, virtual, 2020."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/1583991.1584053"},{"key":"e_1_3_2_1_18_1","volume-title":"8th International Conference on Learning Representations, ICLR 2020","author":"Cai Han","year":"2020","unstructured":"Han Cai , Chuang Gan , Tianzhe Wang , Zhekai Zhang , and Song Han . Once-for-all : Train one network and specialize it for efficient deployment . In 8th International Conference on Learning Representations, ICLR 2020 , Addis Ababa, Ethiopia, April 26--30 , 2020 . OpenReview.net, 2020. Han Cai, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. Once-for-all: Train one network and specialize it for efficient deployment. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26--30, 2020. OpenReview.net, 2020."},{"key":"e_1_3_2_1_19_1","first-page":"12270","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Chen Minghao","year":"2021","unstructured":"Minghao Chen , Houwen Peng , Jianlong Fu , and Haibin Ling . Autoformer : Searching transformers for visual recognition . In Proceedings of the IEEE\/CVF International Conference on Computer Vision , pages 12270 -- 12280 , 2021 . Minghao Chen, Houwen Peng, Jianlong Fu, and Haibin Ling. Autoformer: Searching transformers for visual recognition. In Proceedings of the IEEE\/CVF International Conference on Computer Vision, pages 12270--12280, 2021."},{"key":"e_1_3_2_1_20_1","first-page":"578","volume-title":"13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18)","author":"Chen Tianqi","year":"2018","unstructured":"Tianqi Chen , Thierry Moreau , Ziheng Jiang , Lianmin Zheng , Eddie Yan , Haichen Shen , Meghan Cowan , Leyuan Wang , Yuwei Hu , Luis Ceze , : An automated {End-to-End} optimizing compiler for deep learning . In 13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18) , pages 578 -- 594 , 2018 . Tianqi Chen, Thierry Moreau, Ziheng Jiang, Lianmin Zheng, Eddie Yan, Haichen Shen, Meghan Cowan, Leyuan Wang, Yuwei Hu, Luis Ceze, et al. {TVM}: An automated {End-to-End} optimizing compiler for deep learning. In 13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18), pages 578--594, 2018."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-2097"},{"key":"e_1_3_2_1_22_1","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","volume":"1","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . BERT : Pre-training of deep bidirectional transformers for language understanding . In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies , Volume 1 (Long and Short Papers), pages 4171--4186, Minneapolis, Minnesota , 2019 . Association for Computational Linguistics. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171--4186, Minneapolis, Minnesota, 2019. Association for Computational Linguistics."},{"key":"e_1_3_2_1_23_1","volume-title":"9th International Conference on Learning Representations, ICLR 2021","author":"Dosovitskiy Alexey","year":"2021","unstructured":"Alexey Dosovitskiy , Lucas Beyer , Alexander Kolesnikov , Dirk Weissenborn , Xiaohua Zhai , Thomas Unterthiner , Mostafa Dehghani , Matthias Minderer , Georg Heigold , Sylvain Gelly , Jakob Uszkoreit , and Neil Houlsby . An image is worth 16x16 words: Transformers for image recognition at scale . In 9th International Conference on Learning Representations, ICLR 2021 , Virtual Event, Austria, May 3--7 , 2021 . OpenReview.net, 2021. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3--7, 2021. OpenReview.net, 2021."},{"key":"e_1_3_2_1_24_1","first-page":"5547","volume-title":"International Conference on Machine Learning","author":"Du Nan","year":"2022","unstructured":"Nan Du , Yanping Huang , Andrew M Dai , Simon Tong , Dmitry Lepikhin , Yuanzhong Xu , Maxim Krikun , Yanqi Zhou , Adams Wei Yu , Orhan Firat , : Efficient scaling of language models with mixture-of-experts . In International Conference on Machine Learning , pages 5547 -- 5569 . PMLR, 2022 . Nan Du, Yanping Huang, Andrew M Dai, Simon Tong, Dmitry Lepikhin, Yuanzhong Xu, Maxim Krikun, Yanqi Zhou, Adams Wei Yu, Orhan Firat, et al. Glam: Efficient scaling of language models with mixture-of-experts. In International Conference on Machine Learning, pages 5547--5569. PMLR, 2022."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1002\/andp.19053221004"},{"key":"e_1_3_2_1_26_1","first-page":"2943","volume-title":"Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research","author":"Evci Utku","year":"2020","unstructured":"Utku Evci , Trevor Gale , Jacob Menick , Pablo Samuel Castro , and Erich Elsen . Rigging the lottery: Making all tickets winners . In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research , pages 2943 -- 2952 . PMLR, 2020 . Utku Evci, Trevor Gale, Jacob Menick, Pablo Samuel Castro, and Erich Elsen. Rigging the lottery: Making all tickets winners. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 2943--2952. PMLR, 2020."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1102"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3437801.3441578"},{"issue":"120","key":"e_1_3_2_1_29_1","first-page":"1","article-title":"Scaling to trillion parameter models with simple and efficient sparsity","volume":"23","author":"Fedus William","year":"2022","unstructured":"William Fedus , Barret Zoph , and Noam Shazeer . Switch transformers : Scaling to trillion parameter models with simple and efficient sparsity . Journal of Machine Learning Research , 23 ( 120 ): 1 -- 39 , 2022 . William Fedus, Barret Zoph, and Noam Shazeer. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. Journal of Machine Learning Research, 23(120):1--39, 2022.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_30_1","first-page":"2023","article-title":"Efficient sparse training with mixture-of-experts","author":"Gale Trevor","year":"2023","unstructured":"Trevor Gale , Deepak Narayanan , Cliff Young , and Matei Zaharia . Megablocks : Efficient sparse training with mixture-of-experts . ML-Sys 2023 , 2023 . Trevor Gale, Deepak Narayanan, Cliff Young, and Matei Zaharia. Megablocks: Efficient sparse training with mixture-of-experts. ML-Sys2023, 2023.","journal-title":"ML-Sys"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/SC41405.2020.00021"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/SC41405.2020.00021"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00272"},{"key":"e_1_3_2_1_34_1","article-title":"Dynamic neural networks: A survey","author":"Han Yizeng","year":"2021","unstructured":"Yizeng Han , Gao Huang , Shiji Song , Le Yang , Honghui Wang , and Yulin Wang . Dynamic neural networks: A survey . IEEE Transactions on Pattern Analysis and Machine Intelligence , 2021 . Yizeng Han, Gao Huang, Shiji Song, Le Yang, Honghui Wang, and Yulin Wang. Dynamic neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3293883.3295712"},{"key":"e_1_3_2_1_37_1","volume-title":"et al. Tutel: Adaptive mixture-of-experts at scale. ArXiv preprint, abs\/2206.03382","author":"Hwang Changho","year":"2022","unstructured":"Changho Hwang , Wei Cui , Yifan Xiong , Ziyue Yang , Ze Liu , Han Hu , Zilong Wang , Rafael Salas , Jithin Jose , Prabhat Ram , et al. Tutel: Adaptive mixture-of-experts at scale. ArXiv preprint, abs\/2206.03382 , 2022 . Changho Hwang, Wei Cui, Yifan Xiong, Ziyue Yang, Ze Liu, Han Hu, Zilong Wang, Rafael Salas, Jithin Jose, Prabhat Ram, et al. Tutel: Adaptive mixture-of-experts at scale. ArXiv preprint, abs\/2206.03382, 2022."},{"key":"e_1_3_2_1_38_1","volume-title":"8th International Conference on Learning Representations, ICLR 2020","author":"Kitaev Nikita","year":"2020","unstructured":"Nikita Kitaev , Lukasz Kaiser , and Anselm Levskaya . Reformer : The efficient transformer . In 8th International Conference on Learning Representations, ICLR 2020 , Addis Ababa, Ethiopia, April 26--30 , 2020 . OpenReview.net, 2020. Nikita Kitaev, Lukasz Kaiser, and Anselm Levskaya. Reformer: The efficient transformer. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26--30, 2020. OpenReview.net, 2020."},{"key":"e_1_3_2_1_39_1","volume-title":"Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles (SOSP 23)","author":"Kwon Woosuk","year":"2023","unstructured":"Woosuk Kwon , Zhuohan Li , Siyuan Zhuang , Ying Shen , Lianmin Zheng , Cody Yu , Joey Gonzalez , Hao Zhang , and Ion Stoica . Efficient memory management for large language model serving with pagedattention . In Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles (SOSP 23) , 2023 . Woosuk Kwon, Zhuohan Li, Siyuan Zhuang, Ying Shen, Lianmin Zheng, Cody Yu, Joey Gonzalez, Hao Zhang, and Ion Stoica. Efficient memory management for large language model serving with pagedattention. In Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles (SOSP 23), 2023."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.829"},{"key":"e_1_3_2_1_41_1","series-title":"Proceedings of Machine Learning Research","first-page":"3744","volume-title":"Proceedings of the 36th International Conference on Machine Learning, ICML","author":"Lee Juho","year":"2019","unstructured":"Juho Lee , Yoonho Lee , Jungtaek Kim , Adam R. Kosiorek , Seungjin Choi , and Yee Whye Teh . Set transformer: A framework for attention-based permutation-invariant neural networks . In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019 , 9--15 June 2019, Long Beach, California, USA , volume 97 of Proceedings of Machine Learning Research , pages 3744 -- 3753 . PMLR , 2019. Juho Lee, Yoonho Lee, Jungtaek Kim, Adam R. Kosiorek, Seungjin Choi, and Yee Whye Teh. Set transformer: A framework for attention-based permutation-invariant neural networks. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9--15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pages 3744--3753. PMLR, 2019."},{"key":"e_1_3_2_1_42_1","volume-title":"9th International Conference on Learning Representations, ICLR 2021","author":"Lepikhin Dmitry","year":"2021","unstructured":"Dmitry Lepikhin , HyoukJoong Lee , Yuanzhong Xu , Dehao Chen , Orhan Firat , Yanping Huang , Maxim Krikun , Noam Shazeer , and Zhifeng Chen . Gshard : Scaling giant models with conditional computation and automatic sharding . In 9th International Conference on Learning Representations, ICLR 2021 , Virtual Event, Austria, May 3--7 , 2021 . OpenReview.net, 2021. Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, Yanping Huang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen. Gshard: Scaling giant models with conditional computation and automatic sharding. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3--7, 2021. OpenReview.net, 2021."},{"key":"e_1_3_2_1_43_1","first-page":"5958","volume-title":"Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research","author":"Li Zhuohan","year":"2020","unstructured":"Zhuohan Li , Eric Wallace , Sheng Shen , Kevin Lin , Kurt Keutzer , Dan Klein , and Joey Gonzalez . Train big, then compress: Rethinking model size for efficient training and inference of transformers . In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research , pages 5958 -- 5968 . PMLR, 2020 . Zhuohan Li, Eric Wallace, Sheng Shen, Kevin Lin, Kurt Keutzer, Dan Klein, and Joey Gonzalez. Train big, then compress: Rethinking model size for efficient training and inference of transformers. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 5958--5968. PMLR, 2020."},{"key":"e_1_3_2_1_44_1","volume-title":"Sashank J Reddi, Ke Ye, Felix Chern, Felix Yu, Ruiqi Guo, et al. Large models are parsimonious learners: Activation sparsity in trained transformers. ArXiv preprint, abs\/2210.06313","author":"Li Zonglin","year":"2022","unstructured":"Zonglin Li , Chong You , Srinadh Bhojanapalli , Daliang Li , Ankit Singh Rawat , Sashank J Reddi, Ke Ye, Felix Chern, Felix Yu, Ruiqi Guo, et al. Large models are parsimonious learners: Activation sparsity in trained transformers. ArXiv preprint, abs\/2210.06313 , 2022 . Zonglin Li, Chong You, Srinadh Bhojanapalli, Daliang Li, Ankit Singh Rawat, Sashank J Reddi, Ke Ye, Felix Chern, Felix Yu, Ruiqi Guo, et al. Large models are parsimonious learners: Activation sparsity in trained transformers. ArXiv preprint, abs\/2210.06313, 2022."},{"key":"e_1_3_2_1_45_1","volume-title":"Transformer acceleration with dynamic sparse attention. ArXiv preprint, abs\/2110.11299","author":"Liu Liu","year":"2021","unstructured":"Liu Liu , Zheng Qu , Zhaodong Chen , Yufei Ding , and Yuan Xie . Transformer acceleration with dynamic sparse attention. ArXiv preprint, abs\/2110.11299 , 2021 . Liu Liu, Zheng Qu, Zhaodong Chen, Yufei Ding, and Yuan Xie. Transformer acceleration with dynamic sparse attention. ArXiv preprint, abs\/2110.11299, 2021."},{"key":"e_1_3_2_1_46_1","volume-title":"Ten lessons we have learned in the new \"sparseland\": A short handbook for sparse neural network researchers. ArXiv, abs\/2302.02596","author":"Liu Shiwei","year":"2023","unstructured":"Shiwei Liu and Zhangyang Wang . Ten lessons we have learned in the new \"sparseland\": A short handbook for sparse neural network researchers. ArXiv, abs\/2302.02596 , 2023 . Shiwei Liu and Zhangyang Wang. Ten lessons we have learned in the new \"sparseland\": A short handbook for sparse neural network researchers. ArXiv, abs\/2302.02596, 2023."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.648"},{"key":"e_1_3_2_1_48_1","first-page":"881","volume-title":"Proceedings of the 14th USENIX Conference on Operating Systems Design and Implementation","author":"Ma Lingxiao","year":"2020","unstructured":"Lingxiao Ma , Zhiqiang Xie , Zhi Yang , Jilong Xue , Youshan Miao , Wei Cui , Wenxiang Hu , Fan Yang , Lintao Zhang , and Lidong Zhou . Rammer : Enabling holistic deep learning compiler optimizations with rtasks . In Proceedings of the 14th USENIX Conference on Operating Systems Design and Implementation , pages 881 -- 897 , 2020 . Lingxiao Ma, Zhiqiang Xie, Zhi Yang, Jilong Xue, Youshan Miao, Wei Cui, Wenxiang Hu, Fan Yang, Lintao Zhang, and Lidong Zhou. Rammer: Enabling holistic deep learning compiler optimizations with rtasks. In Proceedings of the 14th USENIX Conference on Operating Systems Design and Implementation, pages 881--897, 2020."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.5555\/2002472.2002491"},{"key":"e_1_3_2_1_50_1","volume-title":"Gpt-4 technical report. ArXiv, abs\/2303.08774","author":"AI.","year":"2023","unstructured":"Open AI. Gpt-4 technical report. ArXiv, abs\/2303.08774 , 2023 . OpenAI. Gpt-4 technical report. ArXiv, abs\/2303.08774, 2023."},{"key":"e_1_3_2_1_51_1","volume-title":"Training language models to follow instructions with human feedback. ArXiv preprint, abs\/2203.02155","author":"Ouyang Long","year":"2022","unstructured":"Long Ouyang , Jeff Wu , Xu Jiang , Diogo Almeida , Carroll L Wainwright , Pamela Mishkin , Chong Zhang , Sandhini Agarwal , Katarina Slama , Alex Ray , Training language models to follow instructions with human feedback. ArXiv preprint, abs\/2203.02155 , 2022 . Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. ArXiv preprint, abs\/2203.02155, 2022."},{"key":"e_1_3_2_1_52_1","first-page":"8024","volume-title":"Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , Luca Antiga , Alban Desmaison , Andreas K\u00f6pf , Edward Yang , Zachary DeVito , Martin Raison , Alykhan Tejani , Sasank Chilamkurthy , Benoit Steiner , Lu Fang , Junjie Bai , and Soumith Chintala . Pytorch : An imperative style, high-performance deep learning library. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d'Alch\u00e9-Buc, Emily B. Fox, and Roman Garnett, editors , Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019 , NeurIPS 2019 , December 8--14, 2019, Vancouver, BC, Canada, pages 8024 -- 8035 , 2019. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas K\u00f6pf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d'Alch\u00e9-Buc, Emily B. Fox, and Roman Garnett, editors, Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8--14, 2019, Vancouver, BC, Canada, pages 8024--8035, 2019."},{"key":"e_1_3_2_1_53_1","first-page":"8748","volume-title":"Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18--24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research","author":"Radford Alec","year":"2021","unstructured":"Alec Radford , Jong Wook Kim , Chris Hallacy , Aditya Ramesh , Gabriel Goh , Sandhini Agarwal , Girish Sastry , Amanda Askell , Pamela Mishkin , Jack Clark , Gretchen Krueger , and Ilya Sutskever . Learning transferable visual models from natural language supervision. In Marina Meila and Tong Zhang, editors , Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18--24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research , pages 8748 -- 8763 . PMLR, 2021 . Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. Learning transferable visual models from natural language supervision. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18--24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pages 8748--8763. PMLR, 2021."},{"key":"e_1_3_2_1_54_1","volume-title":"Learning-based methods for comparing sequences, with applications to audio-to-midi alignment and matching","author":"Raffel Colin","year":"2016","unstructured":"Colin Raffel . Learning-based methods for comparing sequences, with applications to audio-to-midi alignment and matching . 2016 . Colin Raffel. Learning-based methods for comparing sequences, with applications to audio-to-midi alignment and matching. 2016."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_2_1_56_1","volume-title":"Hierarchical text-conditional image generation with clip latents. ArXiv preprint, abs\/2204.06125","author":"Ramesh Aditya","year":"2022","unstructured":"Aditya Ramesh , Prafulla Dhariwal , Alex Nichol , Casey Chu , and Mark Chen . Hierarchical text-conditional image generation with clip latents. ArXiv preprint, abs\/2204.06125 , 2022 . Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. ArXiv preprint, abs\/2204.06125, 2022."},{"key":"e_1_3_2_1_57_1","volume-title":"Dynamicvit: Efficient vision transformers with dynamic token sparsification. Advances in neural information processing systems, 34:13937--13949","author":"Rao Yongming","year":"2021","unstructured":"Yongming Rao , Wenliang Zhao , Benlin Liu , Jiwen Lu , Jie Zhou , and Cho-Jui Hsieh . Dynamicvit: Efficient vision transformers with dynamic token sparsification. Advances in neural information processing systems, 34:13937--13949 , 2021 . Yongming Rao, Wenliang Zhao, Benlin Liu, Jiwen Lu, Jie Zhou, and Cho-Jui Hsieh. Dynamicvit: Efficient vision transformers with dynamic token sparsification. Advances in neural information processing systems, 34:13937--13949, 2021."},{"key":"e_1_3_2_1_58_1","volume-title":"Stanford alpaca: An instruction-following llama model. https:\/\/github.com\/tatsu-lab\/stanford_alpaca","author":"Taori Rohan","year":"2023","unstructured":"Rohan Taori , Ishaan Gulrajani , Tianyi Zhang , Yann Dubois , Xuechen Li , Carlos Guestrin , Percy Liang , and Tatsunori B. Hashimoto . Stanford alpaca: An instruction-following llama model. https:\/\/github.com\/tatsu-lab\/stanford_alpaca , 2023 . Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, and Tatsunori B. Hashimoto. Stanford alpaca: An instruction-following llama model. https:\/\/github.com\/tatsu-lab\/stanford_alpaca, 2023."},{"key":"e_1_3_2_1_59_1","volume-title":"7th International Conference on Learning Representations, ICLR 2019","author":"Wang Alex","year":"2019","unstructured":"Alex Wang , Amanpreet Singh , Julian Michael , Felix Hill , Omer Levy , and Samuel R. Bowman . GLUE: A multi-task benchmark and analysis platform for natural language understanding . In 7th International Conference on Learning Representations, ICLR 2019 , New Orleans, LA, USA, May 6--9 , 2019 . OpenReview.net, 2019. Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6--9, 2019. OpenReview.net, 2019."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA51647.2021.00018"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3410463.3414654"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/1362622.1362674"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_64_1","volume-title":"Sparsetir: Composable abstractions for sparse compilation in deep learning. ArXiv preprint, abs\/2207.04606","author":"Ye Zihao","year":"2022","unstructured":"Zihao Ye , Ruihang Lai , Junru Shao , Tianqi Chen , and Luis Ceze . Sparsetir: Composable abstractions for sparse compilation in deep learning. ArXiv preprint, abs\/2207.04606 , 2022 . Zihao Ye, Ruihang Lai, Junru Shao, Tianqi Chen, and Luis Ceze. Sparsetir: Composable abstractions for sparse compilation in deep learning. ArXiv preprint, abs\/2207.04606, 2022."},{"key":"e_1_3_2_1_65_1","volume-title":"Alice H. Oh","author":"Yu Botao","year":"2022","unstructured":"Botao Yu , Peiling Lu , Rui Wang , Wei Hu , Xu Tan , Wei Ye , Shikun Zhang , Tao Qin , and Tie-Yan Liu . Museformer: Transformer with fine- and coarse-grained attention for music generation . In Alice H. Oh , Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems , 2022 . Botao Yu, Peiling Lu, Rui Wang, Wei Hu, Xu Tan, Wei Ye, Shikun Zhang, Tao Qin, and Tie-Yan Liu. Museformer: Transformer with fine- and coarse-grained attention for music generation. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022."},{"key":"e_1_3_2_1_66_1","volume-title":"Xi Victoria Lin, et al. Opt: Open pre-trained transformer language models. ArXiv preprint, abs\/2205.01068","author":"Zhang Susan","year":"2022","unstructured":"Susan Zhang , Stephen Roller , Naman Goyal , Mikel Artetxe , Moya Chen , Shuohui Chen , Christopher Dewan , Mona Diab , Xian Li , Xi Victoria Lin, et al. Opt: Open pre-trained transformer language models. ArXiv preprint, abs\/2205.01068 , 2022 . Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, et al. Opt: Open pre-trained transformer language models. ArXiv preprint, abs\/2205.01068, 2022."},{"key":"e_1_3_2_1_67_1","first-page":"863","volume-title":"14th USENIX symposium on operating systems design and implementation (OSDI 20)","author":"Zheng Lianmin","year":"2020","unstructured":"Lianmin Zheng , Chengfan Jia , Minmin Sun , Zhao Wu , Cody Hao Yu , Ameer Haj-Ali , Yida Wang , Jun Yang , Danyang Zhuo , Koushik Sen , : Generating {High-Performance} tensor programs for deep learning . In 14th USENIX symposium on operating systems design and implementation (OSDI 20) , pages 863 -- 879 , 2020 . Lianmin Zheng, Chengfan Jia, Minmin Sun, Zhao Wu, Cody Hao Yu, Ameer Haj-Ali, Yida Wang, Jun Yang, Danyang Zhuo, Koushik Sen, et al. Ansor: Generating {High-Performance} tensor programs for deep learning. In 14th USENIX symposium on operating systems design and implementation (OSDI 20), pages 863--879, 2020."},{"key":"e_1_3_2_1_68_1","first-page":"213","volume-title":"16th USENIX Symposium on Operating Systems Design and Implementation (OSDI 22)","author":"Zheng Ningxin","year":"2022","unstructured":"Ningxin Zheng , Bin Lin , Quanlu Zhang , Lingxiao Ma , Yuqing Yang , Fan Yang , Yang Wang , Mao Yang , and Lidong Zhou . {SparTA}:{Deep-Learning} model sparsity via {Tensor-with-Sparsity-Attribute} . In 16th USENIX Symposium on Operating Systems Design and Implementation (OSDI 22) , pages 213 -- 232 , 2022 . Ningxin Zheng, Bin Lin, Quanlu Zhang, Lingxiao Ma, Yuqing Yang, Fan Yang, Yang Wang, Mao Yang, and Lidong Zhou. {SparTA}:{Deep-Learning} model sparsity via {Tensor-with-Sparsity-Attribute}. In 16th USENIX Symposium on Operating Systems Design and Implementation (OSDI 22), pages 213--232, 2022."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"e_1_3_2_1_70_1","article-title":"Towards efficient acceleration of transformers using dynamic sparse attention","author":"Zhou Zhe","year":"2022","unstructured":"Zhe Zhou , Junlin Liu , Zhenyu Gu , and Guangyu Sun . Energon : Towards efficient acceleration of transformers using dynamic sparse attention . IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems , 2022 . Zhe Zhou, Junlin Liu, Zhenyu Gu, and Guangyu Sun. Energon: Towards efficient acceleration of transformers using dynamic sparse attention. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 2022.","journal-title":"IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems"},{"key":"e_1_3_2_1_71_1","first-page":"233","volume-title":"16th USENIX Symposium on Operating Systems Design and Implementation (OSDI 22)","author":"Zhu Hongyu","year":"2022","unstructured":"Hongyu Zhu , Ruofan Wu , Yijia Diao , Shanbin Ke , Haoyu Li , Chen Zhang , Jilong Xue , Lingxiao Ma , Yuqing Xia , Wei Cui , Fan Yang , Mao Yang , Lidong Zhou , Asaf Cidon , and Gennady Pekhimenko . ROLLER : Fast and efficient tensor compilation for deep learning . In 16th USENIX Symposium on Operating Systems Design and Implementation (OSDI 22) , pages 233 -- 248 , Carlsbad, CA , 2022 . USENIX Association. Hongyu Zhu, Ruofan Wu, Yijia Diao, Shanbin Ke, Haoyu Li, Chen Zhang, Jilong Xue, Lingxiao Ma, Yuqing Xia, Wei Cui, Fan Yang, Mao Yang, Lidong Zhou, Asaf Cidon, and Gennady Pekhimenko. ROLLER: Fast and efficient tensor compilation for deep learning. In 16th USENIX Symposium on Operating Systems Design and Implementation (OSDI 22), pages 233--248, Carlsbad, CA, 2022. USENIX Association."},{"key":"e_1_3_2_1_72_1","volume-title":"St-moe: Designing stable and transferable sparse expert models. ArXiv preprint, abs\/2202.08906","author":"Zoph Barret","year":"2022","unstructured":"Barret Zoph , Irwan Bello , Sameer Kumar , Nan Du , Yanping Huang , Jeff Dean , Noam Shazeer , and William Fedus . St-moe: Designing stable and transferable sparse expert models. ArXiv preprint, abs\/2202.08906 , 2022 . Barret Zoph, Irwan Bello, Sameer Kumar, Nan Du, Yanping Huang, Jeff Dean, Noam Shazeer, and William Fedus. St-moe: Designing stable and transferable sparse expert models. ArXiv preprint, abs\/2202.08906, 2022."}],"event":{"name":"SOSP '23: 29th Symposium on Operating Systems Principles","location":"Koblenz Germany","acronym":"SOSP '23","sponsor":["SIGOPS ACM Special Interest Group on Operating Systems","USENIX"]},"container-title":["Proceedings of the 29th Symposium on Operating Systems Principles"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3600006.3613139","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:49Z","timestamp":1750178209000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3600006.3613139"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,23]]},"references-count":72,"alternative-id":["10.1145\/3600006.3613139","10.1145\/3600006"],"URL":"https:\/\/doi.org\/10.1145\/3600006.3613139","relation":{},"subject":[],"published":{"date-parts":[[2023,10,23]]},"assertion":[{"value":"2023-10-23","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}