{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,18]],"date-time":"2025-11-18T09:34:41Z","timestamp":1763458481562,"version":"build-2065373602"},"reference-count":59,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,9,8]]},"DOI":"10.1109\/mlcad65511.2025.11189152","type":"proceedings-article","created":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T17:51:22Z","timestamp":1760032282000},"page":"1-10","source":"Crossref","is-referenced-by-count":1,"title":["SALAD: Systematic Assessment of Machine Unlearning on LLM-Aided Hardware Design"],"prefix":"10.1109","author":[{"given":"Zeng","family":"Wang","sequence":"first","affiliation":[{"name":"NYU Tandon School of Engineering,USA"}]},{"given":"Minghao","family":"Shao","sequence":"additional","affiliation":[{"name":"NYU Tandon School of Engineering,USA"}]},{"given":"Rupesh Raj","family":"Karn","sequence":"additional","affiliation":[{"name":"NYU,Abu Dhabi,UAE"}]},{"given":"Likhitha","family":"Mankali","sequence":"additional","affiliation":[{"name":"NYU Tandon School of Engineering,USA"}]},{"given":"Jitendra","family":"Bhandari","sequence":"additional","affiliation":[{"name":"NYU Tandon School of Engineering,USA"}]},{"given":"Ramesh","family":"Karri","sequence":"additional","affiliation":[{"name":"NYU Tandon School of Engineering,USA"}]},{"given":"Ozgur","family":"Sinanoglu","sequence":"additional","affiliation":[{"name":"NYU,Abu Dhabi,UAE"}]},{"given":"Muhammad","family":"Shafique","sequence":"additional","affiliation":[{"name":"NYU,Abu Dhabi,UAE"}]},{"given":"Johann","family":"Knechtel","sequence":"additional","affiliation":[{"name":"NYU,Abu Dhabi,UAE"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"crossref","DOI":"10.1109\/LAD62341.2024.10691788","article-title":"Rtlcoder: Outperforming gpt-3.5 in design rtl generation with our open-source dataset and lightweight solution","author":"Liu","year":"2024"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1145\/3643681"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICLAD65226.2025.00017"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICLAD65226.2025.00018"},{"key":"ref5","first-page":"1","article-title":"Verilogeval: Evaluating large language models for verilog code generation","volume-title":"2023 ICCAD","author":"Liu"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ASP-DAC58780.2024.10473904"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2024.3482107"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ISVLSI61997.2024.00076"},{"article-title":"Au-tochip: Automating hdl generation using llm feedback","year":"2023","author":"Thakur","key":"ref9"},{"article-title":"Llm-assisted generation of hardware assertions","year":"2023","author":"Kande","key":"ref10"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/LAD62341.2024.10691792"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/MLCAD62225.2024.10740250"},{"article-title":"Llm-aided testbench generation and bug detection for finite-state machines","year":"2024","author":"Bhandari","key":"ref13"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TCAD.2024.3383347"},{"article-title":"Chipnemo: Domain-adapted llms for chip design","year":"2023","author":"Liu","key":"ref15"},{"article-title":"Llama 2: Open foundation and fine-tuned chat models","year":"2023","author":"Touvron","key":"ref16"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/MLCAD58807.2023.10299874"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCAD57390.2023.10323953"},{"article-title":"Chipgpt: How far are we from natural language hardware design","year":"2023","author":"Chang","key":"ref19"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.23919\/DATE64628.2025.10993072"},{"article-title":"Craftrtl: High-quality synthetic data generation for verilog code models with correct-by-construction non-textual representations and targeted code repair","year":"2024","author":"Liu","key":"ref21"},{"key":"ref22","article-title":"Deeprtl: Bridging verilog understanding and generation with a unified representation model","author":"Liu","year":"2025","journal-title":"preprint arXiv:2502.15832"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/dac63849.2025.11133191"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3676536.3676830"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1145\/3610721"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18293\/seke2022-066"},{"key":"ref27","first-page":"40 373","article-title":"Codeipprompt: intellectual property infringement assessment of code language models","volume-title":"ICML","author":"Yu"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-96-8183-9_25"},{"key":"ref29","first-page":"2633","article-title":"Extracting training data from large language models","author":"Carlini","year":"2021","journal-title":"USENIX Security 21"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512225"},{"key":"ref31","first-page":"2133","article-title":"{CodexLeaks}: Privacy leaks from code generation language models in {GitHub} copilot","volume-title":"USENIX Security 23","author":"Niu"},{"key":"ref32","first-page":"1559","article-title":"You autocomplete me: Poisoning vulnerabilities in neural code completion","author":"Schuster","year":"2021","journal-title":"USENIX Security 21"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2024.3367788"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1145\/3658644.3690279"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-short.129"},{"article-title":"The emerged security and privacy of llm agent: A survey with case studies","year":"2024","author":"He","key":"ref36"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CCWC62904.2025.10903912"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.23919\/date64628.2025.10993260"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-025-00985-0"},{"key":"ref40","first-page":"12 581","article-title":"Reversing the forget-retain objectives: An efficient llm unlearning framework from logit difference","volume-title":"NeurIPS","author":"Ji"},{"key":"ref41","first-page":"118 198","article-title":"Large language model unlearning via embedding-corrupted prompts","author":"Liu","year":"2024","journal-title":"NeurIPS"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-emnlp.630"},{"article-title":"Llm unlearning via loss adjustment with only forget data","year":"2024","author":"Wang","key":"ref43"},{"article-title":"Tofu: A task of fictitious unlearning for llms","year":"2024","author":"Maini","key":"ref44"},{"article-title":"Negative preference optimization: From catastrophic collapse to effective unlearning","year":"2024","author":"Zhang","key":"ref45"},{"article-title":"Simplicity prevails: Rethinking negative preference optimization for llm unlearning","year":"2024","author":"Fan","key":"ref46"},{"article-title":"The wmdp benchmark: Measuring and reducing malicious use with unlearning","year":"2024","author":"Li","key":"ref47"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/LAD62341.2024.10691810"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3133985"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/HST.2019.8741036"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TETC.2021.3102832"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.23919\/DATE56975.2023.10137020"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-70947-0_5"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/ISLPED58423.2023.10244324"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/ISQED57927.2023.10129292"},{"key":"ref56","first-page":"74","article-title":"Rouge: A package for automatic evaluation of summaries","author":"Lin","year":"2004","journal-title":"Text summarization branches out"},{"article-title":"Detecting pretraining data from large language models","year":"2023","author":"Shi","key":"ref57"},{"key":"ref58","article-title":"Min-k%++: Improved baseline for detecting pre-training data from large language models","author":"Zhang","year":"2024","journal-title":"preprint arXiv:2404.02936"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-ijcnlp.25"}],"event":{"name":"2025 ACM\/IEEE 7th Symposium on Machine Learning for CAD (MLCAD)","start":{"date-parts":[[2025,9,8]]},"location":"Santa Cruz, CA, USA","end":{"date-parts":[[2025,9,10]]}},"container-title":["2025 ACM\/IEEE 7th Symposium on Machine Learning for CAD (MLCAD)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11189084\/11189054\/11189152.pdf?arnumber=11189152","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T17:35:05Z","timestamp":1760117705000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11189152\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,8]]},"references-count":59,"URL":"https:\/\/doi.org\/10.1109\/mlcad65511.2025.11189152","relation":{},"subject":[],"published":{"date-parts":[[2025,9,8]]}}}