{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T15:19:58Z","timestamp":1774365598523,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":70,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,30]],"date-time":"2024-05-30T00:00:00Z","timestamp":1717027200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100006374","name":"National Science Foundation","doi-asserted-by":"publisher","award":["IIS-2310254, CCF-1937500, CNS-1909172"],"award-info":[{"award-number":["IIS-2310254, CCF-1937500, CNS-1909172"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Army Research Office\/Army Research Laboratory","award":["W911-NF-20-1-0167"],"award-info":[{"award-number":["W911-NF-20-1-0167"]}]},{"name":"NSERC Discovery","award":["RGPIN-2019-04613, DGECR-2019-00120"],"award-info":[{"award-number":["RGPIN-2019-04613, DGECR-2019-00120"]}]},{"name":"CFI John R. Evans Leaders Fund"},{"name":"SRC Artificial Intelligence Hardware program"},{"name":"NSERC Alliance","award":["ALLRP-552042-2020"],"award-info":[{"award-number":["ALLRP-552042-2020"]}]},{"name":"BC Knowledge Development Fund"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,30]]},"DOI":"10.1145\/3650200.3656622","type":"proceedings-article","created":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T14:11:54Z","timestamp":1717423914000},"page":"324-337","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6644-4761","authenticated-orcid":false,"given":"Zhengang","family":"Li","sequence":"first","affiliation":[{"name":"Northeastern University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3315-7368","authenticated-orcid":false,"given":"Alec","family":"Lu","sequence":"additional","affiliation":[{"name":"Simon Fraser University, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4325-521X","authenticated-orcid":false,"given":"Yanyue","family":"Xie","sequence":"additional","affiliation":[{"name":"Northeastern University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8120-4456","authenticated-orcid":false,"given":"Zhenglun","family":"Kong","sequence":"additional","affiliation":[{"name":"Northeastern University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3540-1464","authenticated-orcid":false,"given":"Mengshu","family":"Sun","sequence":"additional","affiliation":[{"name":"Beijing University of Technology, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2077-1246","authenticated-orcid":false,"given":"Hao","family":"Tang","sequence":"additional","affiliation":[{"name":"ETH Zurich, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-7237-0236","authenticated-orcid":false,"given":"Zhong Jia","family":"Xue","sequence":"additional","affiliation":[{"name":"Simon Fraser University, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5287-5149","authenticated-orcid":false,"given":"Peiyan","family":"Dong","sequence":"additional","affiliation":[{"name":"Northeastern University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0891-1231","authenticated-orcid":false,"given":"Caiwen","family":"Ding","sequence":"additional","affiliation":[{"name":"University of Connecticut, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3024-7990","authenticated-orcid":false,"given":"Yanzhi","family":"Wang","sequence":"additional","affiliation":[{"name":"Northeastern University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6210-8883","authenticated-orcid":false,"given":"Xue","family":"Lin","sequence":"additional","affiliation":[{"name":"Northeastern University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0603-9697","authenticated-orcid":false,"given":"Zhenman","family":"Fang","sequence":"additional","affiliation":[{"name":"Simon Fraser University, Canada"}]}],"member":"320","published-online":{"date-parts":[[2024,6,3]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.334"},{"key":"e_1_3_2_1_2_1","volume-title":"International Conference on Learning Representations (ICLR).","author":"Baker Bowen","year":"2017","unstructured":"Bowen Baker, Otkrist Gupta, Nikhil Naik, and Ramesh Raskar. 2017. Designing Neural Network Architectures using Reinforcement Learning. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_3_1","volume-title":"Proceedings of the International Conference on Machine Learning (ICML). 550\u2013559","author":"Bender Gabriel","year":"2018","unstructured":"Gabriel Bender, Pieter-Jan Kindermans, Barret Zoph, Vijay Vasudevan, and Quoc Le. 2018. Understanding and simplifying one-shot architecture search. In Proceedings of the International Conference on Machine Learning (ICML). 550\u2013559."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11709"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA51647.2021.00027"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01205"},{"key":"e_1_3_2_1_7_1","volume-title":"Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan.","author":"Choi Jungwook","year":"2018","unstructured":"Jungwook Choi, Zhuo Wang, Swagath Venkataramani, Pierce I-Jen Chuang, Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan. 2018. Pact: Parameterized clipping activation for quantized neural networks. arXiv:1805.06085 (2018)."},{"key":"e_1_3_2_1_8_1","volume-title":"Binaryconnect: Training deep neural networks with binary weights during propagations. In Advances in Neural Information Processing Systems (NeurIPS). 3123\u20133131.","author":"Courbariaux Matthieu","year":"2015","unstructured":"Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David. 2015. Binaryconnect: Training deep neural networks with binary weights during propagations. In Advances in Neural Information Processing Systems (NeurIPS). 3123\u20133131."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00038"},{"key":"e_1_3_2_1_10_1","volume-title":"International Conference on Learning Representations (ICLR). https:\/\/openreview.net\/forum?id=YicbFdNTTy","author":"Dosovitskiy Alexey","year":"2021","unstructured":"Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations (ICLR). https:\/\/openreview.net\/forum?id=YicbFdNTTy"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58517-4_32"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01170"},{"key":"e_1_3_2_1_14_1","volume-title":"Rethinking spatial dimensions of vision transformers. arXiv:2103.16302","author":"Heo Byeongho","year":"2021","unstructured":"Byeongho Heo, Sangdoo Yun, Dongyoon Han, Sanghyuk Chun, Junsuk Choe, and Seong\u00a0Joon Oh. 2021. Rethinking spatial dimensions of vision transformers. arXiv:2103.16302 (2021)."},{"key":"e_1_3_2_1_15_1","volume-title":"Distilling the knowledge in a neural network. arXiv:1503.02531","author":"Hinton Geoffrey","year":"2015","unstructured":"Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_16_1","volume-title":"Intel\u00ae Xeon\u00ae Silver 4214 Processor. https:\/\/ark.intel.com\/content\/www\/us\/en\/ark\/products\/193385\/intel-xeon-silver-4214-processor-16-5m-cache-2-20-ghz.html. Last accessed","year":"2024","unstructured":"Intel. 2024. Intel\u00ae Xeon\u00ae Silver 4214 Processor. https:\/\/ark.intel.com\/content\/www\/us\/en\/ark\/products\/193385\/intel-xeon-silver-4214-processor-16-5m-cache-2-20-ghz.html. Last accessed Jan. 18, 2024."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11713"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01206"},{"key":"e_1_3_2_1_19_1","unstructured":"Yuhang Li Xin Dong and Wei Wang. 2020. Additive Powers-of-Two Quantization: An Efficient Non-uniform Discretization for Neural Networks. In International Conference on Learning Representations (ICLR). https:\/\/openreview.net\/forum?id=BkgXT24tDS"},{"key":"e_1_3_2_1_20_1","unstructured":"Yanjing Li Sheng Xu Baochang Zhang Xianbin Cao Peng Gao and Guodong Guo. 2022. Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer. In Advances in Neural Information Processing Systems (NeurIPS). https:\/\/openreview.net\/forum?id=fU-m9kQe0ke"},{"key":"e_1_3_2_1_21_1","unstructured":"Yawei Li Kai Zhang Jiezhang Cao Radu Timofte and Luc\u00a0Van Gool. 2021. LocalViT: Bringing Locality to Vision Transformers. arxiv:2104.05707\u00a0[cs.CV]"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/FPL57034.2022.00027"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01403"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/164"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_2"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"e_1_3_2_1_27_1","unstructured":"Zhenhua Liu Yunhe Wang Kai Han Siwei Ma and Wen Gao. 2021. Post-Training Quantization for Vision Transformer. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_28_1","volume-title":"Decoupled Weight Decay Regularization. In International Conference on Learning Representations (ICLR). https:\/\/openreview.net\/forum?id=Bkg6RiCqY7","author":"Loshchilov Ilya","year":"2019","unstructured":"Ilya Loshchilov and Frank Hutter. 2019. Decoupled Weight Decay Regularization. In International Conference on Learning Representations (ICLR). https:\/\/openreview.net\/forum?id=Bkg6RiCqY7"},{"key":"e_1_3_2_1_29_1","volume-title":"AutoQ: Automated Kernel-Wise Neural Network Quantization. In International Conference on Learning Representations (ICLR).","author":"Lou Qian","year":"2019","unstructured":"Qian Lou, Feng Guo, Minje Kim, Lantao Liu, and Lei Jiang. 2019. AutoQ: Automated Kernel-Wise Neural Network Quantization. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_30_1","volume-title":"Artificial Intelligence in the Age of Neural Networks and Brain Computing","author":"Miikkulainen Risto","unstructured":"Risto Miikkulainen, Jason Liang, Elliot Meyerson, Aditya Rawal, Dan Fink, Olivier Francon, Bala Raju, Hormoz Shahrzad, Arshak Navruzyan, Nigel Duffy, 2019. Evolving deep neural networks. In Artificial Intelligence in the Age of Neural Networks and Brain Computing. Elsevier, 293\u2013312."},{"key":"e_1_3_2_1_31_1","first-page":"24898","article-title":"IA-RED2: Interpretability-Aware Redundancy Reduction for Vision Transformers","volume":"34","author":"Pan Bowen","year":"2021","unstructured":"Bowen Pan, Rameswar Panda, Yifan Jiang, Zhangyang Wang, Rogerio Feris, and Aude Oliva. 2021. IA-RED2: Interpretability-Aware Redundancy Reduction for Vision Transformers. Advances in Neural Information Processing Systems (NeurIPS) 34 (2021), 24898\u201324911.","journal-title":"Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"e_1_3_2_1_32_1","volume-title":"Proceedings of the International Conference on Machine Learning (ICML). 4095\u20134104","author":"Pham Hieu","year":"2018","unstructured":"Hieu Pham, Melody Guan, Barret Zoph, Quoc Le, and Jeff Dean. 2018. Efficient neural architecture search via parameters sharing. In Proceedings of the International Conference on Machine Learning (ICML). 4095\u20134104."},{"key":"e_1_3_2_1_33_1","volume-title":"https:\/\/pytorch.org\/tutorials\/recipes\/recipes\/profiler_recipe.html. Last accessed","author":"PYTORCH","year":"2024","unstructured":"PyTorch. 2024. PYTORCH PROFILER. https:\/\/pytorch.org\/tutorials\/recipes\/recipes\/profiler_recipe.html. Last accessed Jan. 18, 2024."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01044"},{"key":"e_1_3_2_1_35_1","unstructured":"Maithra Raghu Thomas Unterthiner Simon Kornblith Chiyuan Zhang and Alexey Dosovitskiy. 2021. Do Vision Transformers See Like Convolutional Neural Networks?. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33014780"},{"key":"e_1_3_2_1_38_1","volume-title":"CompOFA: Compound Once-For-All Networks for Faster Multi-Platform Deployment. arXiv preprint arXiv:2104.12642","author":"Sahni Manas","year":"2021","unstructured":"Manas Sahni, Shreya Varshini, Alind Khare, and Alexey Tumanov. 2021. CompOFA: Compound Once-For-All Networks for Faster Multi-Platform Deployment. arXiv preprint arXiv:2104.12642 (2021)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6409"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490422.3502364"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00293"},{"key":"e_1_3_2_1_42_1","volume-title":"Proceedings of the International Conference on Machine Learning (ICML). PMLR, 10347\u201310357","author":"Touvron Hugo","year":"2021","unstructured":"Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herv\u00e9 J\u00e9gou. 2021. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning (ICML). PMLR, 10347\u201310357."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00010"},{"key":"e_1_3_2_1_44_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_45_1","volume-title":"International Conference on Learning Representations (ICLR).","author":"Uhlich Stefan","year":"2020","unstructured":"Stefan Uhlich, Lukas Mauch, Fabien Cardinaux, Kazuki Yoshiyama, Javier\u00a0Alonso Garcia, Stephen Tiedemann, Thomas Kemp, and Akira Nakamura. 2020. Mixed Precision DNNs: All you need is a good parametrization. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_46_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems (NeurIPS). 5998\u20136008."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00635"},{"key":"e_1_3_2_1_48_1","volume-title":"Hat: Hardware-aware transformers for efficient natural language processing. arXiv:2005.14187","author":"Wang Hanrui","year":"2020","unstructured":"Hanrui Wang, Zhanghao Wu, Zhijian Liu, Han Cai, Ligeng Zhu, Chuang Gan, and Song Han. 2020. Hat: Hardware-aware transformers for efficient natural language processing. arXiv:2005.14187 (2020)."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00881"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00215"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503221.3508408"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01099"},{"key":"e_1_3_2_1_54_1","volume-title":"Mixed precision quantization of convnets via differentiable neural architecture search. arXiv:1812.00090","author":"Wu Bichen","year":"2018","unstructured":"Bichen Wu, Yanghan Wang, Peizhao Zhang, Yuandong Tian, Peter Vajda, and Kurt Keutzer. 2018. Mixed precision quantization of convnets via differentiable neural architecture search. arXiv:1812.00090 (2018)."},{"key":"e_1_3_2_1_55_1","volume-title":"Cvt: Introducing convolutions to vision transformers. arXiv preprint arXiv:2103.15808","author":"Wu Haiping","year":"2021","unstructured":"Haiping Wu, Bin Xiao, Noel Codella, Mengchen Liu, Xiyang Dai, Lu Yuan, and Lei Zhang. 2021. Cvt: Introducing convolutions to vision transformers. arXiv preprint arXiv:2103.15808 (2021)."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.634"},{"key":"e_1_3_2_1_57_1","volume-title":"Deep Learning with INT8 Optimization on Xilinx Devices. https:\/\/docs.xilinx.com\/v\/u\/en-US\/wp486-deep-learning-int8. Last accessed","year":"2022","unstructured":"Xilinx. 2017. Deep Learning with INT8 Optimization on Xilinx Devices. https:\/\/docs.xilinx.com\/v\/u\/en-US\/wp486-deep-learning-int8. Last accessed Mar. 28, 2022."},{"key":"e_1_3_2_1_58_1","volume-title":"Convolutional Neural Network with INT4 Optimization on Xilinx Devices. https:\/\/docs.xilinx.com\/v\/u\/en-US\/wp521-4bit-optimization. Last accessed","year":"2022","unstructured":"Xilinx. 2020. Convolutional Neural Network with INT4 Optimization on Xilinx Devices. https:\/\/docs.xilinx.com\/v\/u\/en-US\/wp521-4bit-optimization. Last accessed Mar. 28, 2022."},{"key":"e_1_3_2_1_59_1","volume-title":"Vivado Design Suite. https:\/\/www.xilinx.com\/products\/design-tools\/vivado.html. Last accessed","year":"2022","unstructured":"Xilinx. 2020. Vivado Design Suite. https:\/\/www.xilinx.com\/products\/design-tools\/vivado.html. Last accessed Aug. 28, 2022."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00207"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58571-6_41"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/EMC2-NIPS53020.2019.00016"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.37"},{"key":"e_1_3_2_1_66_1","volume-title":"Proceedings of the International Conference on Machine Learning (ICML). PMLR, 12707\u201312718","author":"Zhao Yiyang","year":"2021","unstructured":"Yiyang Zhao, Linnan Wang, Yuandong Tian, Rodrigo Fonseca, and Tian Guo. 2021. Few-shot neural architecture search. In Proceedings of the International Conference on Machine Learning (ICML). PMLR, 12707\u201312718."},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00257"},{"key":"e_1_3_2_1_68_1","volume-title":"Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv:1606.06160","author":"Zhou Shuchang","year":"2016","unstructured":"Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, and Yuheng Zou. 2016. Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv:1606.06160 (2016)."},{"key":"e_1_3_2_1_69_1","volume-title":"Neural Architecture Search with Reinforcement Learning. In International Conference on Learning Representations (ICLR).","author":"Zoph Barret","year":"2017","unstructured":"Barret Zoph and Quoc\u00a0V. Le. 2017. Neural Architecture Search with Reinforcement Learning. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00907"}],"event":{"name":"ICS '24: 2024 International Conference on Supercomputing","location":"Kyoto Japan","acronym":"ICS '24","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture"]},"container-title":["Proceedings of the 38th ACM International Conference on Supercomputing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3650200.3656622","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3650200.3656622","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T15:22:50Z","timestamp":1755876170000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3650200.3656622"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,30]]},"references-count":70,"alternative-id":["10.1145\/3650200.3656622","10.1145\/3650200"],"URL":"https:\/\/doi.org\/10.1145\/3650200.3656622","relation":{},"subject":[],"published":{"date-parts":[[2024,5,30]]},"assertion":[{"value":"2024-06-03","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}