{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T12:11:04Z","timestamp":1767183064945,"version":"3.37.3"},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T00:00:00Z","timestamp":1635724800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T00:00:00Z","timestamp":1635724800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T00:00:00Z","timestamp":1635724800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,11,1]]},"DOI":"10.1109\/iccad51958.2021.9643471","type":"proceedings-article","created":{"date-parts":[[2021,12,23]],"date-time":"2021-12-23T23:06:46Z","timestamp":1640300806000},"page":"1-9","source":"Crossref","is-referenced-by-count":4,"title":["A Framework for Area-efficient Multi-task BERT Execution on ReRAM-based Accelerators"],"prefix":"10.1109","author":[{"given":"Myeonggu","family":"Kang","sequence":"first","affiliation":[]},{"given":"Hyein","family":"Shin","sequence":"additional","affiliation":[]},{"given":"Jaekang","family":"Shin","sequence":"additional","affiliation":[]},{"given":"Lee-Sup","family":"Kim","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"journal-title":"Masking as an efficient alternative to finetuning for pretrained language models","year":"2020","author":"zhao","key":"ref30"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1612"},{"journal-title":"Multi-task deep neural networks for natural language understanding","year":"2019","author":"liu","key":"ref11"},{"key":"ref12","article-title":"Roberta: A robustly optimized BERT pretraining approach","volume":"abs 1907 11692","author":"liu","year":"2019","journal-title":"CoRR"},{"key":"ref13","doi-asserted-by":"crossref","first-page":"1769","DOI":"10.23919\/DATE.2019.8715178","article-title":"Design of reliable dnn accelerator with un-reliable reram","author":"long","year":"2019","journal-title":"2019 Design Automation Test in Europe Conference Exhibition (DATE)"},{"journal-title":"Achieving real-time execution of transformer-based large-scale models on mobile with compiler-aware neural architecture optimization","year":"2020","author":"niu","key":"ref14"},{"journal-title":"A Survey of the Usages of Deep Learning in Natural Language Processing","year":"2018","author":"otter","key":"ref15"},{"key":"ref16","first-page":"8024","article-title":"Pytorch: An imperative style, high-performance deep learning library","author":"paszke","year":"2019","journal-title":"Proc NeurIPS"},{"journal-title":"Performance analysis of deep learning workloads on leading-edge systems","year":"2019","author":"ren","key":"ref17"},{"journal-title":"On the effect of dropping layers of pre-trained transformer models","year":"2021","author":"sajjad","key":"ref18"},{"journal-title":"DistilBERT a distilled version of BERT smaller faster cheaper and lighter","year":"2019","author":"sanh","key":"ref19"},{"journal-title":"Deebert Dynamic early exiting for accelerating bert inference","year":"2020","author":"xin","key":"ref28"},{"key":"ref4","article-title":"BERT: pre-training of deep bidirectional transformers for language understanding","volume":"abs 1810 4805","author":"devlin","year":"2018","journal-title":"CoRR"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/TCAD.2017.2729466"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3007787.3001140"},{"journal-title":"Power-bert Accelerating bert inference via progressive word-vector elimination","year":"2020","author":"goyal","key":"ref6"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/3400302.3415640"},{"journal-title":"Compressing BERT Studying the effects of weight pruning on transfer learning","year":"2020","author":"gordon","key":"ref5"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/MM.2011.42"},{"journal-title":"Parameter-efficient transfer learning for NLP","year":"2019","author":"houlsby","key":"ref7"},{"key":"ref2","article-title":"Analysis methods in neural language processing: A survey","volume":"abs 1812 8951","author":"belinkov","year":"2018","journal-title":"CoRR"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2002.1017616"},{"key":"ref1","first-page":"5151","article-title":"Scalable methods for 8-bit training of neural networks","author":"banner","year":"2018","journal-title":"Proceedings of the 32nd International Conference on Neural Information Processing Systems ser NIPS'18"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3007787.3001139"},{"journal-title":"MobileBERT a Compact Task-Agnostic BERT for Resource-Limited Devices","year":"2020","author":"sun","key":"ref22"},{"key":"ref21","article-title":"Matching the blanks: Distributional similarity for relation learning","volume":"abs 1906 3158","author":"soares","year":"2019","journal-title":"CoRR"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5446"},{"journal-title":"Sequence to sequence learning with neural networks","year":"2014","author":"sutskever","key":"ref23"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/216585.216588"},{"journal-title":"HuggingFace's Transformers State-of-the-art Natural Language Processing","year":"2019","author":"wolf","key":"ref25"}],"event":{"name":"2021 IEEE\/ACM International Conference On Computer Aided Design (ICCAD)","start":{"date-parts":[[2021,11,1]]},"location":"Munich, Germany","end":{"date-parts":[[2021,11,4]]}},"container-title":["2021 IEEE\/ACM International Conference On Computer Aided Design (ICCAD)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9643423\/9643432\/09643471.pdf?arnumber=9643471","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T16:55:15Z","timestamp":1652201715000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9643471\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,1]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/iccad51958.2021.9643471","relation":{},"subject":[],"published":{"date-parts":[[2021,11,1]]}}}