{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:03:39Z","timestamp":1775815419627,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":70,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,12,19]],"date-time":"2022-12-19T00:00:00Z","timestamp":1671408000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,12,19]]},"DOI":"10.1145\/3567955.3567961","type":"proceedings-article","created":{"date-parts":[[2022,12,21]],"date-time":"2022-12-21T18:24:44Z","timestamp":1671647084000},"page":"123-137","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":17,"title":["TelaMalloc: Efficient On-Chip Memory Allocation for Production Machine Learning Accelerators"],"prefix":"10.1145","author":[{"given":"Martin","family":"Maas","sequence":"first","affiliation":[{"name":"Google, USA"}]},{"given":"Ulysse","family":"Beaugnon","sequence":"additional","affiliation":[{"name":"Google, France"}]},{"given":"Arun","family":"Chauhan","sequence":"additional","affiliation":[{"name":"Google, USA"}]},{"given":"Berkin","family":"Ilbeyi","sequence":"additional","affiliation":[{"name":"Google, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,12,21]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2017. TensorFlow Lite. https:\/\/www.tensorflow.org\/lite \t\t\t\t  2017. TensorFlow Lite. https:\/\/www.tensorflow.org\/lite"},{"key":"e_1_3_2_1_2_1","unstructured":"2020. Optimizing TensorFlow Lite Runtime Memory. https:\/\/blog.tensorflow.org\/2020\/10\/optimizing-tensorflow-lite-runtime.html \t\t\t\t  2020. Optimizing TensorFlow Lite Runtime Memory. https:\/\/blog.tensorflow.org\/2020\/10\/optimizing-tensorflow-lite-runtime.html"},{"key":"e_1_3_2_1_3_1","unstructured":"2021. Google OR Tools: CP-SAT Solver. https:\/\/developers.google.com\/optimization\/cp\/cp_solver \t\t\t\t  2021. Google OR Tools: CP-SAT Solver. https:\/\/developers.google.com\/optimization\/cp\/cp_solver"},{"key":"e_1_3_2_1_4_1","unstructured":"2021. Google Tensor is a milestone for machine learning. https:\/\/blog.google\/products\/pixel\/introducing-google-tensor\/ \t\t\t\t  2021. Google Tensor is a milestone for machine learning. https:\/\/blog.google\/products\/pixel\/introducing-google-tensor\/"},{"key":"e_1_3_2_1_5_1","unstructured":"2022. Android Neural Network API. https:\/\/developer.android.com\/ndk\/guides\/neuralnetworks \t\t\t\t  2022. Android Neural Network API. https:\/\/developer.android.com\/ndk\/guides\/neuralnetworks"},{"key":"e_1_3_2_1_6_1","unstructured":"2022. pprof. https:\/\/github.com\/google\/pprof \t\t\t\t  2022. pprof. https:\/\/github.com\/google\/pprof"},{"key":"e_1_3_2_1_7_1","unstructured":"2022. TensorFlow GitHub Repository: BFC Allocator. https:\/\/github.com\/tensorflow\/tensorflow\/blob\/master\/tensorflow\/core\/common_runtime\/bfc_allocator.h \t\t\t\t  2022. TensorFlow GitHub Repository: BFC Allocator. https:\/\/github.com\/tensorflow\/tensorflow\/blob\/master\/tensorflow\/core\/common_runtime\/bfc_allocator.h"},{"key":"e_1_3_2_1_8_1","unstructured":"2022. TensorFlow GitHub Repository: Memory Repacker. https:\/\/github.com\/tensorflow\/tensorflow\/blob\/master\/tensorflow\/compiler\/xla\/service\/memory_space_assignment_repacking.h \t\t\t\t  2022. TensorFlow GitHub Repository: Memory Repacker. https:\/\/github.com\/tensorflow\/tensorflow\/blob\/master\/tensorflow\/compiler\/xla\/service\/memory_space_assignment_repacking.h"},{"key":"e_1_3_2_1_9_1","unstructured":"2022. Yggdrasil Decision Forests. https:\/\/github.com\/google\/yggdrasil-decision-forests \t\t\t\t  2022. Yggdrasil Decision Forests. https:\/\/github.com\/google\/yggdrasil-decision-forests"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306346.3322967"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW56347.2022.00300"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/774789.774805"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3033019.3033023"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-00142-0_69"},{"key":"e_1_3_2_1_15_1","volume-title":"Chris Leary, Dougal Maclaurin, George Necula, Adam Paszke, Jake VanderPlas, Skye Wanderman-Milne, and Qiao Zhang.","author":"Bradbury James","year":"2018","unstructured":"James Bradbury , Roy Frostig , Peter Hawkins , Matthew James Johnson , Chris Leary, Dougal Maclaurin, George Necula, Adam Paszke, Jake VanderPlas, Skye Wanderman-Milne, and Qiao Zhang. 2018 . JAX: composable transformations of Python +NumPy programs. http:\/\/github.com\/google\/jax James Bradbury, Roy Frostig, Peter Hawkins, Matthew James Johnson, Chris Leary, Dougal Maclaurin, George Necula, Adam Paszke, Jake VanderPlas, Skye Wanderman-Milne, and Qiao Zhang. 2018. JAX: composable transformations of Python+NumPy programs. http:\/\/github.com\/google\/jax"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2929257"},{"key":"e_1_3_2_1_17_1","unstructured":"Quentin Cappart Thierry Moisan Louis-Martin Rousseau Isabeau Pr\u00e9mont-Schwarz and Andre Cire. 2020. Combining Reinforcement Learning and Constraint Programming for Combinatorial Optimization. arxiv:2006.01610. \t\t\t\t  Quentin Cappart Thierry Moisan Louis-Martin Rousseau Isabeau Pr\u00e9mont-Schwarz and Andre Cire. 2020. Combining Reinforcement Learning and Constraint Programming for Combinatorial Optimization. arxiv:2006.01610."},{"key":"e_1_3_2_1_18_1","volume-title":"Efficient Data-Parallel Pipelines. In ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI). http:\/\/dl.acm.org\/citation.cfm?id=1806638","author":"Chambers Craig","year":"2010","unstructured":"Craig Chambers , Ashish Raniwala , Frances Perry , Stephen Adams , Robert Henry , Robert Bradshaw , and Nathan. 2010 . FlumeJava: Easy , Efficient Data-Parallel Pipelines. In ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI). http:\/\/dl.acm.org\/citation.cfm?id=1806638 Craig Chambers, Ashish Raniwala, Frances Perry, Stephen Adams, Robert Henry, Robert Bradshaw, and Nathan. 2010. FlumeJava: Easy, Efficient Data-Parallel Pipelines. In ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI). http:\/\/dl.acm.org\/citation.cfm?id=1806638"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485137"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2014.6853198"},{"key":"e_1_3_2_1_21_1","volume-title":"TVM: An Automated End-to-End Optimizing Compiler for Deep Learning. In 13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18)","author":"Chen Tianqi","year":"2018","unstructured":"Tianqi Chen , Thierry Moreau , Ziheng Jiang , Lianmin Zheng , Eddie Yan , Haichen Shen , Meghan Cowan , Leyuan Wang , Yuwei Hu , Luis Ceze , Carlos Guestrin , and Arvind Krishnamurthy . 2018 . TVM: An Automated End-to-End Optimizing Compiler for Deep Learning. In 13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18) . isbn:978-1-939133-08-3 https:\/\/www.usenix.org\/conference\/osdi18\/presentation\/chen Tianqi Chen, Thierry Moreau, Ziheng Jiang, Lianmin Zheng, Eddie Yan, Haichen Shen, Meghan Cowan, Leyuan Wang, Yuwei Hu, Luis Ceze, Carlos Guestrin, and Arvind Krishnamurthy. 2018. TVM: An Automated End-to-End Optimizing Compiler for Deep Learning. In 13th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18). isbn:978-1-939133-08-3 https:\/\/www.usenix.org\/conference\/osdi18\/presentation\/chen"},{"key":"e_1_3_2_1_22_1","volume-title":"Training Deep Nets with Sublinear Memory Cost. CoRR, abs\/1604.06174","author":"Chen Tianqi","year":"2016","unstructured":"Tianqi Chen , Bing Xu , Chiyuan Zhang , and Carlos Guestrin . 2016. Training Deep Nets with Sublinear Memory Cost. CoRR, abs\/1604.06174 ( 2016 ), arXiv:1604.06174. arxiv:1604.06174 Tianqi Chen, Bing Xu, Chiyuan Zhang, and Carlos Guestrin. 2016. Training Deep Nets with Sublinear Memory Cost. CoRR, abs\/1604.06174 (2016), arXiv:1604.06174. arxiv:1604.06174"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.5555\/3327144.3327258"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSSC.2016.2616357"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/FPL.2018.00009"},{"key":"e_1_3_2_1_26_1","volume-title":"Tcmalloc: Thread-caching malloc.","author":"Ghemawat Sanjay","year":"2009","unstructured":"Sanjay Ghemawat and Paul Menage . 2009 . Tcmalloc: Thread-caching malloc. Sanjay Ghemawat and Paul Menage. 2009. Tcmalloc: Thread-caching malloc."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3464298.3476132"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3445814.3446762"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA52012.2021.00050"},{"key":"e_1_3_2_1_30_1","volume-title":"Proceedings of Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V. Sze (Eds.). 2, 497\u2013511","author":"Jain Paras","year":"2020","unstructured":"Paras Jain , Ajay Jain , Aniruddha Nrusimha , Amir Gholami , Pieter Abbeel , Joseph Gonzalez , Kurt Keutzer , and Ion Stoica . 2020 . Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization . In Proceedings of Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V. Sze (Eds.). 2, 497\u2013511 . https:\/\/proceedings.mlsys.org\/paper\/2020\/file\/084b6fbb10729ed4da8c3d3f5a3ae7c9-Paper.pdf Paras Jain, Ajay Jain, Aniruddha Nrusimha, Amir Gholami, Pieter Abbeel, Joseph Gonzalez, Kurt Keutzer, and Ion Stoica. 2020. Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization. In Proceedings of Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V. Sze (Eds.). 2, 497\u2013511. https:\/\/proceedings.mlsys.org\/paper\/2020\/file\/084b6fbb10729ed4da8c3d3f5a3ae7c9-Paper.pdf"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA52012.2021.00010"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3400302.3415639"},{"key":"e_1_3_2_1_33_1","volume-title":"Proceedings of Machine Learning and Systems, A. Smola, A. Dimakis, and I. Stoica (Eds.). 3, 387\u2013400","author":"Kaufman Sam","year":"2021","unstructured":"Sam Kaufman , Phitchaya Phothilimthana , Yanqi Zhou , Charith Mendis , Sudip Roy , Amit Sabne , and Mike Burrows . 2021 . A Learned Performance Model for Tensor Processing Units . In Proceedings of Machine Learning and Systems, A. Smola, A. Dimakis, and I. Stoica (Eds.). 3, 387\u2013400 . https:\/\/proceedings.mlsys.org\/paper\/2021\/file\/85d8ce590ad8981ca2c8286f79f59954-Paper.pdf Sam Kaufman, Phitchaya Phothilimthana, Yanqi Zhou, Charith Mendis, Sudip Roy, Amit Sabne, and Mike Burrows. 2021. A Learned Performance Model for Tensor Processing Units. In Proceedings of Machine Learning and Systems, A. Smola, A. Dimakis, and I. Stoica (Eds.). 3, 387\u2013400. https:\/\/proceedings.mlsys.org\/paper\/2021\/file\/85d8ce590ad8981ca2c8286f79f59954-Paper.pdf"},{"key":"e_1_3_2_1_34_1","unstructured":"Shauharda Khadka Estelle Aflalo Mattias Marder Avrech Ben-David Santiago Miret Shie Mannor Tamir Hazan Hanlin Tang and Somdeb Majumdar. 2020. Optimizing Memory Placement using Evolutionary Graph Reinforcement Learning. arxiv:2007.07298. \t\t\t\t  Shauharda Khadka Estelle Aflalo Mattias Marder Avrech Ben-David Santiago Miret Shie Mannor Tamir Hazan Hanlin Tang and Somdeb Majumdar. 2020. Optimizing Memory Placement using Evolutionary Graph Reinforcement Learning. arxiv:2007.07298."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3352460.3358252"},{"key":"e_1_3_2_1_36_1","unstructured":"Doug Lea and Wolfram Gloger. 1996. A memory allocator. \t\t\t\t  Doug Lea and Wolfram Gloger. 1996. A memory allocator."},{"key":"e_1_3_2_1_37_1","volume-title":"XLA: TensorFlow, compiled. TensorFlow Dev Summit.","author":"Leary Chris","year":"2017","unstructured":"Chris Leary and Todd Wang . 2017 . XLA: TensorFlow, compiled. TensorFlow Dev Summit. Chris Leary and Todd Wang. 2017. XLA: TensorFlow, compiled. TensorFlow Dev Summit."},{"key":"e_1_3_2_1_38_1","volume-title":"Efficient Memory Management for Deep Neural Net Inference. In MLSys 2020 Workshop on Resource-Constrained Machine Learning (ReCoML","author":"Lee Juhyun","year":"2020","unstructured":"Juhyun Lee and Yury Pisarchyk . 2020 . Efficient Memory Management for Deep Neural Net Inference. In MLSys 2020 Workshop on Resource-Constrained Machine Learning (ReCoML 2020). Juhyun Lee and Yury Pisarchyk. 2020. Efficient Memory Management for Deep Neural Net Inference. In MLSys 2020 Workshop on Resource-Constrained Machine Learning (ReCoML 2020)."},{"key":"e_1_3_2_1_39_1","volume-title":"GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=qrwe7XHTmYb","author":"Lepikhin Dmitry","year":"2021","unstructured":"Dmitry Lepikhin , HyoukJoong Lee , Yuanzhong Xu , Dehao Chen , Orhan Firat , Yanping Huang , Maxim Krikun , Noam Shazeer , and Zhifeng Chen . 2021 . GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=qrwe7XHTmYb Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, Yanping Huang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen. 2021. GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=qrwe7XHTmYb"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3339861"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/2694344.2694385"},{"key":"e_1_3_2_1_42_1","unstructured":"Changxi Liu Hailong Yang Rujun Sun Zhongzhi Luan Lin Gan Guangwen Yang and Depei Qian. 2019. swTVM: Exploring the Automated Compilation for Deep Learning on Sunway Architecture. arxiv:1904.07404. \t\t\t\t  Changxi Liu Hailong Yang Rujun Sun Zhongzhi Luan Lin Gan Guangwen Yang and Depei Qian. 2019. swTVM: Exploring the Automated Compilation for Deep Learning on Sunway Architecture. arxiv:1904.07404."},{"key":"e_1_3_2_1_43_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning, Hal Daum\u00e9 III and Aarti Singh (Eds.) (Proceedings of Machine Learning Research","volume":"6247","author":"Liu Evan","year":"2020","unstructured":"Evan Liu , Milad Hashemi , Kevin Swersky , Parthasarathy Ranganathan , and Junwhan Ahn . 2020 . An Imitation Learning Approach for Cache Replacement . In Proceedings of the 37th International Conference on Machine Learning, Hal Daum\u00e9 III and Aarti Singh (Eds.) (Proceedings of Machine Learning Research , Vol. 119). PMLR, 6237\u2013 6247 . http:\/\/proceedings.mlr.press\/v119\/liu20f.html Evan Liu, Milad Hashemi, Kevin Swersky, Parthasarathy Ranganathan, and Junwhan Ahn. 2020. An Imitation Learning Approach for Cache Replacement. In Proceedings of the 37th International Conference on Machine Learning, Hal Daum\u00e9 III and Aarti Singh (Eds.) (Proceedings of Machine Learning Research, Vol. 119). PMLR, 6237\u20136247. http:\/\/proceedings.mlr.press\/v119\/liu20f.html"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/MM.2020.3012883"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/TC.2021.3059962"},{"key":"e_1_3_2_1_46_1","volume-title":"Hierarchical Planning for Device Placement. In International Conference on Learning Representations. https:\/\/openreview.net\/pdf?id=Hkc-TeZ0W","author":"Mirhoseini Azalia","year":"2018","unstructured":"Azalia Mirhoseini , Anna Goldie , Hieu Pham , Benoit Steiner , Quoc V. Le , and Jeff Dean . 2018 . Hierarchical Planning for Device Placement. In International Conference on Learning Representations. https:\/\/openreview.net\/pdf?id=Hkc-TeZ0W Azalia Mirhoseini, Anna Goldie, Hieu Pham, Benoit Steiner, Quoc V. Le, and Jeff Dean. 2018. Hierarchical Planning for Device Placement. In International Conference on Learning Representations. https:\/\/openreview.net\/pdf?id=Hkc-TeZ0W"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-021-03544-w"},{"key":"e_1_3_2_1_48_1","unstructured":"Vinod Nair Sergey Bartunov Felix Gimeno Ingrid von Glehn Pawel Lichocki Ivan Lobov Brendan O\u2019Donoghue Nicolas Sonnerat Christian Tjandraatmadja Pengming Wang Ravichandra Addanki Tharindi Hapuarachchi Thomas Keck James Keeling Pushmeet Kohli Ira Ktena Yujia Li Oriol Vinyals and Yori Zwols. 2021. Solving Mixed Integer Programs Using Neural Networks. arxiv:2012.13349. \t\t\t\t  Vinod Nair Sergey Bartunov Felix Gimeno Ingrid von Glehn Pawel Lichocki Ivan Lobov Brendan O\u2019Donoghue Nicolas Sonnerat Christian Tjandraatmadja Pengming Wang Ravichandra Addanki Tharindi Hapuarachchi Thomas Keck James Keeling Pushmeet Kohli Ira Ktena Yujia Li Oriol Vinyals and Yori Zwols. 2021. Solving Mixed Integer Programs Using Neural Networks. arxiv:2012.13349."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3243176.3243212"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/2491956.2462163"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISPASS.2019.00042"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/PACT52795.2021.00008"},{"key":"e_1_3_2_1_53_1","volume-title":"2014 International Conference on Hardware\/Software Codesign and System Synthesis (CODES+ISSS).","author":"Pilato Christian","unstructured":"Christian Pilato , Paolo Mantovani , Giuseppe Di Guglielmo , and Luca P. Carloni . 2014. System-level memory optimization for high-level synthesis of component-based SoCs . In 2014 International Conference on Hardware\/Software Codesign and System Synthesis (CODES+ISSS). Christian Pilato, Paolo Mantovani, Giuseppe Di Guglielmo, and Luca P. Carloni. 2014. System-level memory optimization for high-level synthesis of component-based SoCs. In 2014 International Conference on Hardware\/Software Codesign and System Synthesis (CODES+ISSS)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3012084"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503465"},{"key":"e_1_3_2_1_56_1","volume-title":"Matthew Denton, and Tushar Krishna.","author":"Samajdar Ananda","year":"2021","unstructured":"Ananda Samajdar , Jan Moritz Joseph , Matthew Denton, and Tushar Krishna. 2021 . AIRCHITECT : Learning Custom Architecture Design and Mapping Space . https:\/\/doi.org\/10.48550\/ARXIV.2108.08295 10.48550\/ARXIV.2108.08295 Ananda Samajdar, Jan Moritz Joseph, Matthew Denton, and Tushar Krishna. 2021. AIRCHITECT: Learning Custom Architecture Design and Mapping Space. https:\/\/doi.org\/10.48550\/ARXIV.2108.08295"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/SC.Companion.2012.111"},{"key":"e_1_3_2_1_58_1","unstructured":"Taro Sekiyama Takashi Imamichi Haruki Imai and Rudy Raymond. 2018. Profile-guided memory optimization for deep neural networks. arXiv preprint arXiv:1804.10001. \t\t\t\t  Taro Sekiyama Takashi Imamichi Haruki Imai and Rudy Raymond. 2018. Profile-guided memory optimization for deep neural networks. arXiv preprint arXiv:1804.10001."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA51647.2021.00075"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/2901318.2901355"},{"key":"e_1_3_2_1_61_1","volume-title":"Dynamic Storage Allocation: A Survey and Critical Review","author":"Wilson Paul R.","unstructured":"Paul R. Wilson , Mark S. Johnstone , Michael Neely , and David Boles . 1995. Dynamic Storage Allocation: A Survey and Critical Review . Springer-Verlag , 1\u2013116. Paul R. Wilson, Mark S. Johnstone, Michael Neely, and David Boles. 1995. Dynamic Storage Allocation: A Survey and Critical Review. Springer-Verlag, 1\u2013116."},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA52012.2021.00086"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3373376.3378514"},{"key":"#cr-split#-e_1_3_2_1_64_1.1","unstructured":"Amir Yazdanbakhsh Kiran Seshadri Berkin Akin James Laudon and Ravi Narayanaswami. 2021. An Evaluation of Edge TPU Accelerators for Convolutional Neural Networks. https:\/\/doi.org\/10.48550\/ARXIV.2102.10423 10.48550\/ARXIV.2102.10423"},{"key":"#cr-split#-e_1_3_2_1_64_1.2","unstructured":"Amir Yazdanbakhsh Kiran Seshadri Berkin Akin James Laudon and Ravi Narayanaswami. 2021. An Evaluation of Edge TPU Accelerators for Convolutional Neural Networks. https:\/\/doi.org\/10.48550\/ARXIV.2102.10423"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/2684746.2689060"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503222.3507767"},{"key":"e_1_3_2_1_67_1","unstructured":"Yanqi Zhou Xuanyi Dong Berkin Akin Mingxing Tan Daiyi Peng Tianjian Meng Amir Yazdanbakhsh Da Huang Ravi Narayanaswami and James Laudon. 2021. Rethinking Co-design of Neural Architectures and Hardware Accelerators. arxiv:2102.08619. \t\t\t\t  Yanqi Zhou Xuanyi Dong Berkin Akin Mingxing Tan Daiyi Peng Tianjian Meng Amir Yazdanbakhsh Da Huang Ravi Narayanaswami and James Laudon. 2021. Rethinking Co-design of Neural Architectures and Hardware Accelerators. arxiv:2102.08619."},{"key":"e_1_3_2_1_68_1","volume-title":"Proceedings of Machine Learning and Systems, D. Marculescu, Y. Chi, and C. Wu (Eds.). 4, 141\u2013152","author":"Zhou Yanqi","year":"2022","unstructured":"Yanqi Zhou , Xuanyi Dong , Tianjian Meng , Mingxing Tan , Berkin Akin , Daiyi Peng , Amir Yazdanbakhsh , Da Huang , Ravi Narayanaswami , and James Laudon . 2022 . Towards the Co-design of Neural Networks and Accelerators . In Proceedings of Machine Learning and Systems, D. Marculescu, Y. Chi, and C. Wu (Eds.). 4, 141\u2013152 . https:\/\/proceedings.mlsys.org\/paper\/2022\/file\/31fefc0e570cb3860f2a6d4b38c6490d-Paper.pdf Yanqi Zhou, Xuanyi Dong, Tianjian Meng, Mingxing Tan, Berkin Akin, Daiyi Peng, Amir Yazdanbakhsh, Da Huang, Ravi Narayanaswami, and James Laudon. 2022. Towards the Co-design of Neural Networks and Accelerators. In Proceedings of Machine Learning and Systems, D. Marculescu, Y. Chi, and C. Wu (Eds.). 4, 141\u2013152. https:\/\/proceedings.mlsys.org\/paper\/2022\/file\/31fefc0e570cb3860f2a6d4b38c6490d-Paper.pdf"},{"key":"e_1_3_2_1_69_1","volume-title":"Advances in Neural Information Processing Systems","author":"Zhou Yanqi","year":"2020","unstructured":"Yanqi Zhou , Sudip Roy , Amirali Abdolrashidi , Daniel Wong , Peter Ma , Qiumin Xu , Hanxiao Liu , Phitchaya Phothilimtha , Shen Wang , Anna Goldie , Azalia Mirhoseini , and James Laudon . 2020. Transferable Graph Optimizers for ML Compilers . In Advances in Neural Information Processing Systems , H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.). 33, https:\/\/proceedings.neurips.cc\/paper\/ 2020 \/file\/9f29450d2eb58feb555078bdefe28aa5-Paper.pdf Yanqi Zhou, Sudip Roy, Amirali Abdolrashidi, Daniel Wong, Peter Ma, Qiumin Xu, Hanxiao Liu, Phitchaya Phothilimtha, Shen Wang, Anna Goldie, Azalia Mirhoseini, and James Laudon. 2020. Transferable Graph Optimizers for ML Compilers. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.). 33, https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/9f29450d2eb58feb555078bdefe28aa5-Paper.pdf"}],"event":{"name":"ASPLOS '23: 28th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 1","location":"Vancouver BC Canada","acronym":"ASPLOS '23","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture","SIGOPS ACM Special Interest Group on Operating Systems","SIGPLAN ACM Special Interest Group on Programming Languages"]},"container-title":["Proceedings of the 28th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 1"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3567955.3567961","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3567955.3567961","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T21:26:14Z","timestamp":1750281974000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3567955.3567961"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,19]]},"references-count":70,"alternative-id":["10.1145\/3567955.3567961","10.1145\/3567955"],"URL":"https:\/\/doi.org\/10.1145\/3567955.3567961","relation":{},"subject":[],"published":{"date-parts":[[2022,12,19]]},"assertion":[{"value":"2022-12-21","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}