{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:06:05Z","timestamp":1750309565161,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":45,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,5,25]],"date-time":"2025-05-25T00:00:00Z","timestamp":1748131200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["UKRI256"],"award-info":[{"award-number":["UKRI256"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/V028251\/1"],"award-info":[{"award-number":["EP\/V028251\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/N031768\/1"],"award-info":[{"award-number":["EP\/N031768\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/S030069\/1"],"award-info":[{"award-number":["EP\/S030069\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/X036006\/1"],"award-info":[{"award-number":["EP\/X036006\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,26]]},"DOI":"10.1145\/3728179.3728192","type":"proceedings-article","created":{"date-parts":[[2025,5,15]],"date-time":"2025-05-15T14:09:58Z","timestamp":1747318198000},"page":"25-34","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["ResBench: A Resource-Aware Benchmark for LLM-Generated FPGA Designs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0272-9175","authenticated-orcid":false,"given":"Ce","family":"Guo","sequence":"first","affiliation":[{"name":"Department of Computing, Imperial College, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-7418-3607","authenticated-orcid":false,"given":"Tong","family":"Zhao","sequence":"additional","affiliation":[{"name":"Department of Computing, Imperial College, London, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2025,5,25]]},"reference":[{"key":"e_1_3_3_2_2_2","doi-asserted-by":"crossref","unstructured":"Manar Abdelatty Jingxiao Ma and Sherief Reda. 2024. MetRex: A Benchmark for Verilog Code Metric Reasoning Using LLMs.","DOI":"10.1145\/3658617.3697625"},{"key":"e_1_3_3_2_3_2","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. GPT-4 Technical Report."},{"key":"e_1_3_3_2_4_2","unstructured":"Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui Kai Dang Xiaodong Deng Yang Fan Wenbin Ge Yu Han Fei Huang et\u00a0al. 2023. Qwen Technical Report."},{"key":"e_1_3_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1145\/3670474.3685948"},{"key":"e_1_3_3_2_6_2","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared\u00a0D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et\u00a0al. 2020. Language Models are Few-Shot Learners. Advances in Neural Information Processing Systems 33 (2020) 1877\u20131901."},{"key":"e_1_3_3_2_7_2","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde De\u00a0Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman et\u00a0al. 2021. Evaluating Large Language Models Trained on Code."},{"key":"e_1_3_3_2_8_2","unstructured":"Aakanksha Chowdhery Sharan Narang Jacob Devlin Maarten Bosma Gaurav Mishra Adam Roberts Paul Barham Hyung\u00a0Won Chung Charles Sutton Sebastian Gehrmann et\u00a0al. 2023. PaLM: Scaling Language Modeling with Pathways. Journal of Machine Learning Research 24 240 (2023) 1\u2013113."},{"key":"e_1_3_3_2_9_2","unstructured":"Jacob Devlin. 2018. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding."},{"key":"e_1_3_3_2_10_2","unstructured":"Daniel Fried Armen Aghajanyan Jessy Lin Sida Wang Eric Wallace Freda Shi Ruiqi Zhong Wen-tau Yih Luke Zettlemoyer and Mike Lewis. 2022. Incoder: A Generative Model for Code Infilling and Synthesis."},{"key":"e_1_3_3_2_11_2","unstructured":"Mingzhe Gao Jieru Zhao Zhe Lin Wenchao Ding Xiaofeng Hou Yu Feng Chao Li and Minyi Guo. 2024. AutoVCoder: A Systematic Framework for Automated Verilog Code Generation using LLMs."},{"key":"e_1_3_3_2_12_2","unstructured":"Dan Hendrycks Collin Burns Saurav Kadavath Akul Arora Steven Basart Eric Tang Dawn Song and Jacob Steinhardt. 2021. Measuring Mathematical Problem Solving with the Math Dataset."},{"key":"e_1_3_3_2_13_2","unstructured":"Binyuan Hui Jian Yang Zeyu Cui Jiaxi Yang Dayiheng Liu Lei Zhang Tianyu Liu Jiajun Zhang Bowen Yu Keming Lu et\u00a0al. 2024. Qwen2.5-Coder Technical Report."},{"key":"e_1_3_3_2_14_2","unstructured":"Aaron Hurst Adam Lerer Adam\u00a0P Goucher Adam Perelman Aditya Ramesh Aidan Clark AJ Ostrow Akila Welihinda Alan Hayes Alec Radford et\u00a0al. 2024. GPT-4o System Card."},{"key":"e_1_3_3_2_15_2","unstructured":"Aaron Jaech Adam Kalai Adam Lerer Adam Richardson Ahmed El-Kishky Aiden Low Alec Helyar Aleksander Madry Alex Beutel Alex Carney et\u00a0al. 2024. OpenAI O1 System Card."},{"key":"e_1_3_3_2_16_2","unstructured":"Albert\u00a0Q Jiang Alexandre Sablayrolles Arthur Mensch Chris Bamford Devendra\u00a0Singh Chaplot Diego de\u00a0las Casas Florian Bressand Gianna Lengyel Guillaume Lample Lucile Saulnier et\u00a0al. 2023. Mistral 7B."},{"key":"e_1_3_3_2_17_2","unstructured":"Juyong Jiang Fan Wang Jiasi Shen Sungju Kim and Sunghun Kim. 2024. A Survey on Large Language Models for Code Generation. arxiv:https:\/\/arXiv.org\/abs\/2406.00515\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2406.00515"},{"key":"e_1_3_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1109\/LAD62341.2024.10691770"},{"key":"e_1_3_3_2_19_2","first-page":"18319","volume-title":"International Conference on Machine Learning","author":"Lai Yuhang","year":"2023","unstructured":"Yuhang Lai, Chengxi Li, Yiming Wang, Tianyi Zhang, Ruiqi Zhong, Luke Zettlemoyer, Wen-tau Yih, Daniel Fried, Sida Wang, and Tao Yu. 2023. DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation. In International Conference on Machine Learning. PMLR, 18319\u201318345."},{"key":"e_1_3_3_2_20_2","unstructured":"Raymond Li Loubna\u00a0Ben Allal Yangtian Zi Niklas Muennighoff Denis Kocetkov Chenghao Mou Marc Marone Christopher Akiki Jia Li Jenny Chim et\u00a0al. 2023. StarCoder: May the Source Be with You!"},{"key":"e_1_3_3_2_21_2","unstructured":"Jiawei Liu Chunqiu\u00a0Steven Xia Yuyao Wang and Lingming Zhang. 2024. Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCAD57390.2023.10323812"},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1145\/3676536.3697118"},{"key":"e_1_3_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/ASP-DAC58780.2024.10473904"},{"key":"e_1_3_3_2_25_2","unstructured":"Niklas Muennighoff Qian Liu Armel Zebaze Qinkai Zheng Binyuan Hui Terry\u00a0Yue Zhuo Swayam Singh Xiangru Tang Leandro Von\u00a0Werra and Shayne Longpre. 2023. Octopack: Instruction Tuning Code Large Language Models."},{"key":"e_1_3_3_2_26_2","doi-asserted-by":"crossref","unstructured":"Bardia Nadimi and Hao Zheng. 2024. A Multi-Expert Large Language Model Architecture for Verilog Code Generation.","DOI":"10.1109\/LAD62341.2024.10691683"},{"key":"e_1_3_3_2_27_2","unstructured":"Erik Nijkamp Bo Pang Hiroaki Hayashi Lifu Tu Huan Wang Yingbo Zhou Silvio Savarese and Caiming Xiong. 2022. CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis."},{"key":"e_1_3_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3670474.3685956"},{"key":"e_1_3_3_2_29_2","unstructured":"Alec Radford. 2018. Improving Language Understanding by Generative Pre-Training."},{"key":"e_1_3_3_2_30_2","unstructured":"Shuo Ren Daya Guo Shuai Lu Long Zhou Shujie Liu Duyu Tang Neel Sundaresan Ming Zhou Ambrosio Blanco and Shuai Ma. 2020. CodeBLEU: A Method for Automatic Evaluation of Code Synthesis."},{"key":"e_1_3_3_2_31_2","unstructured":"Baptiste Roziere Jonas Gehring Fabian Gloeckle Sten Sootla Itai Gat Xiaoqing\u00a0Ellen Tan Yossi Adi Jingyu Liu Romain Sauvestre Tal Remez et\u00a0al. 2023. Code Llama: Open Foundation Models for Code."},{"key":"e_1_3_3_2_32_2","unstructured":"Alvin Tan. 2017. HDLBits: Digital Circuits Exercises. https:\/\/hdlbits.01xz.net\/wiki\/Problem_sets Accessed: 2025."},{"key":"e_1_3_3_2_33_2","doi-asserted-by":"publisher","DOI":"10.23919\/DATE56975.2023.10137086"},{"key":"e_1_3_3_2_34_2","doi-asserted-by":"crossref","unstructured":"Shailja Thakur Baleegh Ahmad Hammond Pearce Benjamin Tan Brendan Dolan-Gavitt Ramesh Karri and Siddharth Garg. 2024. VeriGen: A Large Language Model for Verilog Code Generation. ACM Transactions on Design Automation of Electronic Systems 29 3 (2024) 1\u201331.","DOI":"10.1145\/3643681"},{"key":"e_1_3_3_2_35_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar et\u00a0al. 2023. LLaMA: Open and Efficient Foundation Language Models."},{"key":"e_1_3_3_2_36_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-emnlp.902"},{"key":"e_1_3_3_2_37_2","unstructured":"Ashish Vaswani. 2017. Attention is All You Need. Advances in Neural Information Processing Systems (2017)."},{"key":"e_1_3_3_2_38_2","doi-asserted-by":"crossref","unstructured":"Prashanth Vijayaraghavan Luyao Shi Stefano Ambrogio Charles Mackin Apoorva Nitsure David Beymer and Ehsan Degan. 2024. VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation.","DOI":"10.1109\/LAD62341.2024.10691836"},{"key":"e_1_3_3_2_39_2","unstructured":"Gwok-Waa Wan Wang yubo SamZaak Wong jingyi zhang Mengnv Xing Zhe jiang Nan Guan ying wang Ning Xu Qiang Xu and Xi Wang. 2025. GenBen:A Genarative Benchmark for LLM-Aided Design. https:\/\/openreview.net\/forum?id=gtVo4xcpFI"},{"key":"e_1_3_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1109\/MedAI59581.2023.00044"},{"key":"e_1_3_3_2_41_2","unstructured":"Ning Wang Bingkun Yao Jie Zhou Xi Wang Zhe Jiang and Nan Guan. 2024. Large Language Model for Verilog Generation with Golden Code Feedback."},{"key":"e_1_3_3_2_42_2","doi-asserted-by":"publisher","DOI":"10.1109\/LAD62341.2024.10691753"},{"key":"e_1_3_3_2_43_2","unstructured":"An Yang Baosong Yang Beichen Zhang Binyuan Hui Bo Zheng Bowen Yu Chengyuan Li Dayiheng Liu Fei Huang Haoran Wei et\u00a0al. 2024. Qwen2.5 Technical Report."},{"key":"e_1_3_3_2_44_2","unstructured":"Yiyao Yang Fu Teng Pengju Liu Mengnan Qi Chenyang Lv Ji Li Xuhong Zhang and Zhezhi He. 2025. HaVen: Hallucination-Mitigated LLM for Verilog Code Generation Aligned with HDL Engineers."},{"key":"e_1_3_3_2_45_2","unstructured":"Junjie Ye Xuanting Chen Nuo Xu Can Zu Zekai Shao Shichun Liu Yuhan Cui Zeyang Zhou Chao Gong Yang Shen et\u00a0al. 2023. A Comprehensive Capability Analysis of GPT-3 and GPT-3.5 Series Models."},{"key":"e_1_3_3_2_46_2","unstructured":"Zibin Zheng Kaiwen Ning Yanlin Wang Jingwen Zhang Dewu Zheng Mingxi Ye and Jiachi Chen. 2023. A Survey of Large Language Models for Code: Evolution Benchmarking and Future Trends."}],"event":{"name":"HEART 2025: 15th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies","acronym":"HEART 2025","location":"Kumamoto Japan"},"container-title":["Proceedings of the 15th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3728179.3728192","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3728179.3728192","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:19:06Z","timestamp":1750295946000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3728179.3728192"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,25]]},"references-count":45,"alternative-id":["10.1145\/3728179.3728192","10.1145\/3728179"],"URL":"https:\/\/doi.org\/10.1145\/3728179.3728192","relation":{},"subject":[],"published":{"date-parts":[[2025,5,25]]},"assertion":[{"value":"2025-05-25","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}