{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,7]],"date-time":"2025-08-07T09:22:08Z","timestamp":1754558528094,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,23]],"date-time":"2024-06-23T00:00:00Z","timestamp":1719100800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,23]]},"DOI":"10.1145\/3649329.3655925","type":"proceedings-article","created":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T19:27:22Z","timestamp":1731007642000},"page":"1-6","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Accelerating DTCO with a Sample-Efficient Active Learning Framework for TCAD Device Modeling"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7321-0269","authenticated-orcid":false,"given":"Chanwoo","family":"Park","sequence":"first","affiliation":[{"name":"Alsemy Inc., Seoul, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-4602-9877","authenticated-orcid":false,"given":"Junghwan","family":"Park","sequence":"additional","affiliation":[{"name":"Alsemy Inc., Seoul, Kyunggi-do, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5002-3978","authenticated-orcid":false,"given":"Premkumar","family":"Vincent","sequence":"additional","affiliation":[{"name":"Alsemy Inc., Seoul, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-0425-5301","authenticated-orcid":false,"given":"Hyunbo","family":"Cho","sequence":"additional","affiliation":[{"name":"Alsemy Inc., Seoul, Seoul, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2024,11,7]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"1","volume-title":"IEEE","author":"Zhang Z.","year":"2019","unstructured":"Z. Zhang et al., \"New-generation design-technology co-optimization (dtco): Machine-learning assisted modeling framework,\" in Silicon Nanoelectronics Workshop (SNW). IEEE, 2019, pp. 1--2."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00976"},{"key":"e_1_3_2_1_3_1","first-page":"1050","article-title":"Dropout as a bayesian approximation: Representing model uncertainty in deep learning","author":"Gal Y.","year":"2016","unstructured":"Y. Gal et al., \"Dropout as a bayesian approximation: Representing model uncertainty in deep learning,\" in Proc. ICML. PMLR, 2016, pp. 1050--1059.","journal-title":"Proc. ICML. PMLR"},{"key":"e_1_3_2_1_4_1","volume-title":"Training-free neural active learning with initialization-robustness guarantees,\" ICML","author":"Hemachandra A.","year":"2023","unstructured":"A. Hemachandra et al., \"Training-free neural active learning with initialization-robustness guarantees,\" ICML, 2023."},{"key":"e_1_3_2_1_5_1","volume-title":"Neural tangent kernel: Convergence and generalization in neural networks,\" NeurIPS","author":"Jacot A.","year":"2018","unstructured":"A. Jacot et al., \"Neural tangent kernel: Convergence and generalization in neural networks,\" NeurIPS, vol. 31, 2018."},{"key":"e_1_3_2_1_6_1","volume-title":"Wide neural networks of any depth evolve as linear models under gradient descent,\" NeurIPS","author":"Lee J.","year":"2019","unstructured":"J. Lee et al., \"Wide neural networks of any depth evolve as linear models under gradient descent,\" NeurIPS, vol. 32, 2019."},{"key":"e_1_3_2_1_7_1","first-page":"257","article-title":"Physics-augmented neural compact model for emerging device technologies","author":"Kim Y.","year":"2020","unstructured":"Y. Kim et al., \"Physics-augmented neural compact model for emerging device technologies,\" in Proc. SISPAD. IEEE, 2020, pp. 257--260.","journal-title":"Proc. SISPAD. IEEE"},{"key":"e_1_3_2_1_8_1","first-page":"359","article-title":"Physics-informed graph neural network for circuit compact model development","author":"Gao X.","year":"2020","unstructured":"X. Gao et al., \"Physics-informed graph neural network for circuit compact model development,\" in Proc. SISPAD. IEEE, 2020, pp. 359--362.","journal-title":"Proc. SISPAD. IEEE"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/TED.2022.3208514"},{"key":"e_1_3_2_1_10_1","volume-title":"Machine learning augmented compact modeling for simultaneous improvement in computational speed and accuracy,\" IEEE TED","author":"Sheelvardhan K.","year":"2023","unstructured":"K. Sheelvardhan et al., \"Machine learning augmented compact modeling for simultaneous improvement in computational speed and accuracy,\" IEEE TED, 2023."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.sse.2022.108500"},{"key":"e_1_3_2_1_12_1","volume-title":"A physical-based artificial neural networks compact modeling framework for emerging fets,\" IEEE TED","author":"Yang Y.-S.","year":"2023","unstructured":"Y.-S. Yang, Y. Li, and S. R. Kola, \"A physical-based artificial neural networks compact modeling framework for emerging fets,\" IEEE TED, 2023."},{"key":"e_1_3_2_1_13_1","volume-title":"A novel prediction technology of output characteristics for igbt based on compact model and artificial neural networks,\" IEEE TED","author":"Yao Q.","year":"2023","unstructured":"Q. Yao et al., \"A novel prediction technology of output characteristics for igbt based on compact model and artificial neural networks,\" IEEE TED, 2023."},{"key":"e_1_3_2_1_14_1","first-page":"1183","article-title":"Deep bayesian active learning with image data","author":"Gal Y.","year":"2017","unstructured":"Y. Gal et al., \"Deep bayesian active learning with image data,\" in Proc. ICML. PMLR, 2017, pp. 1183--1192.","journal-title":"Proc. ICML. PMLR"},{"key":"e_1_3_2_1_15_1","volume-title":"Active learning for convolutional neural networks: A core-set approach,\" arXiv preprint","author":"Sener O.","year":"2017","unstructured":"O. Sener et al., \"Active learning for convolutional neural networks: A core-set approach,\" arXiv preprint, 2017, arXiv:1708.00489."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"S. Natarajan et al. \"A 32nm logic technology featuring 2nd-generation high-k + metal-gate transistors enhanced channel strain and 0.171 \u03bcm2 sram cell size in a 291mb array \" 2008 IEEE International Electron Devices Meeting pp. 1--3 2008.","DOI":"10.1109\/IEDM.2008.4796777"}],"event":{"name":"DAC '24: 61st ACM\/IEEE Design Automation Conference","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE-CEDA","SIGBED ACM Special Interest Group on Embedded Systems"],"location":"San Francisco CA USA","acronym":"DAC '24"},"container-title":["Proceedings of the 61st ACM\/IEEE Design Automation Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3649329.3655925","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3649329.3655925","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:17:48Z","timestamp":1750295868000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3649329.3655925"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,23]]},"references-count":16,"alternative-id":["10.1145\/3649329.3655925","10.1145\/3649329"],"URL":"https:\/\/doi.org\/10.1145\/3649329.3655925","relation":{},"subject":[],"published":{"date-parts":[[2024,6,23]]},"assertion":[{"value":"2024-11-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}