{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T01:36:23Z","timestamp":1768268183745,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":23,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T00:00:00Z","timestamp":1725840000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,9,9]]},"DOI":"10.1145\/3670474.3685966","type":"proceedings-article","created":{"date-parts":[[2024,9,3]],"date-time":"2024-09-03T06:22:27Z","timestamp":1725344547000},"page":"1-10","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["Chain-of-Descriptions: Improving Code LLMs for VHDL Code Generation and Summarization"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9981-3954","authenticated-orcid":false,"given":"Prashanth","family":"Vijayaraghavan","sequence":"first","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-8964-4862","authenticated-orcid":false,"given":"Apoorva","family":"Nitsure","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8413-5583","authenticated-orcid":false,"given":"Charles","family":"Mackin","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4703-3294","authenticated-orcid":false,"given":"Luyao","family":"Shi","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5475-4209","authenticated-orcid":false,"given":"Stefano","family":"Ambrogio","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-3493-6918","authenticated-orcid":false,"given":"Arvind","family":"Haran","sequence":"additional","affiliation":[{"name":"IBM, Austin, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4696-5855","authenticated-orcid":false,"given":"Viresh","family":"Paruthi","sequence":"additional","affiliation":[{"name":"IBM, Austin, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-7718-0476","authenticated-orcid":false,"given":"Ali","family":"Elzein","sequence":"additional","affiliation":[{"name":"IBM, Austin, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-1705-0804","authenticated-orcid":false,"given":"Dan","family":"Coops","sequence":"additional","affiliation":[{"name":"IBM, Austin, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9491-4967","authenticated-orcid":false,"given":"David","family":"Beymer","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-9345-8490","authenticated-orcid":false,"given":"Tyler","family":"Baldwin","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6927-8719","authenticated-orcid":false,"given":"Ehsan","family":"Degan","sequence":"additional","affiliation":[{"name":"IBM Research, San Jose, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,9,9]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_3_2_1_2_1","volume-title":"Proceedings of the First IEEE International Workshop on LLM-Aided Design","author":"Anonymous","year":"2024","unstructured":"Anonymous. 2024. VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation. In Press. Proceedings of the First IEEE International Workshop on LLM-Aided Design (2024)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCD.2006.4380826"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2007.70725"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/MLCAD58807.2023.10299874"},{"key":"e_1_3_2_1_6_1","volume-title":"Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, et al.","author":"Chen Mark","year":"2021","unstructured":"Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, et al. 2021. Evaluating large language models trained on code. arXiv preprint arXiv:2107.03374 (2021)."},{"key":"e_1_3_2_1_7_1","volume-title":"Chain-of-verification reduces hallucination in large language models. arXiv preprint arXiv:2309.11495","author":"Dhuliawala Shehzaad","year":"2023","unstructured":"Shehzaad Dhuliawala, Mojtaba Komeili, Jing Xu, Roberta Raileanu, Xian Li, Asli Celikyilmaz, and Jason Weston. 2023. Chain-of-verification reduces hallucination in large language models. arXiv preprint arXiv:2309.11495 (2023)."},{"key":"e_1_3_2_1_8_1","volume-title":"Yangtian Zi, Niklas Muennighoff, Denis Kocetkov, Chenghao Mou, Marc Marone, Christopher Akiki, Jia Li, Jenny Chim, et al.","author":"Li Raymond","year":"2023","unstructured":"Raymond Li, Loubna Ben Allal, Yangtian Zi, Niklas Muennighoff, Denis Kocetkov, Chenghao Mou, Marc Marone, Christopher Akiki, Jia Li, Jenny Chim, et al. 2023. Starcoder: may the source be with you! arXiv preprint arXiv:2305.06161 (2023)."},{"key":"e_1_3_2_1_9_1","volume-title":"Chipnemo: Domain-adapted llms for chip design. arXiv preprint arXiv:2311.00176","author":"Liu Mingjie","year":"2023","unstructured":"Mingjie Liu, Teodor-Dumitru Ene, Robert Kirby, Chris Cheng, Nathaniel Pinckney, Rongjian Liang, Jonah Alben, Himyanshu Anand, Sanmitra Banerjee, Ismet Bayraktaroglu, et al. 2023. Chipnemo: Domain-adapted llms for chip design. arXiv preprint arXiv:2311.00176 (2023)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCAD57390.2023.10323812"},{"key":"e_1_3_2_1_11_1","volume-title":"Beyond accuracy: Evaluating self-consistency of code large language models with identitychain. arXiv preprint arXiv:2310.14053","author":"Min Marcus J","year":"2023","unstructured":"Marcus J Min, Yangruibo Ding, Luca Buratti, Saurabh Pujar, Gail Kaiser, Suman Jana, and Baishakhi Ray. 2023. Beyond accuracy: Evaluating self-consistency of code large language models with identitychain. arXiv preprint arXiv:2310.14053 (2023)."},{"key":"e_1_3_2_1_12_1","volume-title":"Michele Merler, Parameswaran Selvam, Saptha Surendran, Shivdeep Singh, et al.","author":"Mishra Mayank","year":"2024","unstructured":"Mayank Mishra, Matt Stallone, Gaoyuan Zhang, Yikang Shen, Aditya Prasad, Adriana Meza Soria, Michele Merler, Parameswaran Selvam, Saptha Surendran, Shivdeep Singh, et al. 2024. Granite code models: A family of open foundation models for code intelligence. arXiv preprint arXiv:2405.04324 (2024)."},{"key":"e_1_3_2_1_13_1","volume-title":"Codegen: An open large language model for code with multi-turn program synthesis. arXiv preprint arXiv:2203.13474","author":"Nijkamp Erik","year":"2022","unstructured":"Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, and Caiming Xiong. 2022. Codegen: An open large language model for code with multi-turn program synthesis. arXiv preprint arXiv:2203.13474 (2022)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3380446.3430634"},{"key":"e_1_3_2_1_15_1","volume-title":"Yossi Adi, Jingyu Liu, Tal Remez, J\u00e9r\u00e9my Rapin, et al.","author":"Roziere Baptiste","year":"2023","unstructured":"Baptiste Roziere, Jonas Gehring, Fabian Gloeckle, Sten Sootla, Itai Gat, Xiaoqing Ellen Tan, Yossi Adi, Jingyu Liu, Tal Remez, J\u00e9r\u00e9my Rapin, et al. 2023. Code llama: Open foundation models for code. arXiv preprint arXiv:2308.12950 (2023)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236024.3236026"},{"key":"e_1_3_2_1_17_1","volume-title":"Megatron-lm: Training multi-billion parameter language models using model parallelism. arXiv preprint arXiv:1909.08053","author":"Shoeybi Mohammad","year":"2019","unstructured":"Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. 2019. Megatron-lm: Training multi-billion parameter language models using model parallelism. arXiv preprint arXiv:1909.08053 (2019)."},{"key":"e_1_3_2_1_18_1","volume-title":"Automation & Test in Europe Conference & Exhibition (DATE). IEEE, 1--6.","author":"Thakur Shailja","year":"2023","unstructured":"Shailja Thakur, Baleegh Ahmad, Zhenxing Fan, Hammond Pearce, Benjamin Tan, Ramesh Karri, Brendan Dolan-Gavitt, and Siddharth Garg. 2023. Benchmarking large language models for automated verilog rtl code generation. In 2023 Design, Automation & Test in Europe Conference & Exhibition (DATE). IEEE, 1--6."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3643681"},{"key":"e_1_3_2_1_20_1","volume-title":"Attention is all you need. Advances in neural information processing systems 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_21_1","volume-title":"Aakanksha Chowdhery, and Denny Zhou.","author":"Wang Xuezhi","year":"2022","unstructured":"Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc Le, Ed Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. 2022. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171 (2022)."},{"key":"e_1_3_2_1_22_1","volume-title":"Denny Zhou, et al.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems 35 (2022), 24824--24837."},{"key":"e_1_3_2_1_23_1","volume-title":"Evaluating instruction-tuned large language models on code comprehension and generation. arXiv preprint arXiv:2308.01240","author":"Yuan Zhiqiang","year":"2023","unstructured":"Zhiqiang Yuan, Junwei Liu, Qiancheng Zi, Mingwei Liu, Xin Peng, and Yiling Lou. 2023. Evaluating instruction-tuned large language models on code comprehension and generation. arXiv preprint arXiv:2308.01240 (2023)."}],"event":{"name":"MLCAD '24: 2024 ACM\/IEEE International Symposium on Machine Learning for CAD","location":"Salt Lake City UT USA","acronym":"MLCAD '24","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE CEDA"]},"container-title":["Proceedings of the 2024 ACM\/IEEE International Symposium on Machine Learning for CAD"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3670474.3685966","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3670474.3685966","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T23:44:47Z","timestamp":1755906287000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3670474.3685966"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,9]]},"references-count":23,"alternative-id":["10.1145\/3670474.3685966","10.1145\/3670474"],"URL":"https:\/\/doi.org\/10.1145\/3670474.3685966","relation":{},"subject":[],"published":{"date-parts":[[2024,9,9]]},"assertion":[{"value":"2024-09-09","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}