{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,8,13]],"date-time":"2024-08-13T07:45:47Z","timestamp":1723535147691},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,6,6]]},"DOI":"10.1109\/icassp39728.2021.9413475","type":"proceedings-article","created":{"date-parts":[[2021,5,13]],"date-time":"2021-05-13T19:53:45Z","timestamp":1620935625000},"source":"Crossref","is-referenced-by-count":4,"title":["Adaptable Multi-Domain Language Model for Transformer ASR"],"prefix":"10.1109","author":[{"given":"Taewoo","family":"Lee","sequence":"first","affiliation":[]},{"given":"Min-Joong","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Tae Gyoon","family":"Kang","sequence":"additional","affiliation":[]},{"given":"Seokyeoung","family":"Jung","sequence":"additional","affiliation":[]},{"given":"Minseok","family":"Kwon","sequence":"additional","affiliation":[]},{"given":"Yeona","family":"Hong","sequence":"additional","affiliation":[]},{"given":"Jungin","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Kyoung-Gu","family":"Woo","sequence":"additional","affiliation":[]},{"given":"Ho-Gyeong","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Jiseung","family":"Jeong","sequence":"additional","affiliation":[]},{"given":"Jihyun","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Hosik","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Young Sang","family":"Choi","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","article-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach","author":"liu","year":"2019"},{"key":"ref11","article-title":"XLNet: Generalized Autoregressive Pretraining for Language Understanding","author":"yang","year":"2019","journal-title":"NeurIPS"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1016\/S0079-7421(08)60536-8"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_37"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1611835114"},{"key":"ref15","article-title":"Progressive Neural Networks","author":"rusu","year":"2016"},{"key":"ref16","article-title":"Parameter-Efficient Transfer Learning for NLP","author":"houlsby","year":"2019","journal-title":"ICML"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00847"},{"key":"ref18","article-title":"BERT and PALs: Projected Attention Layers for Efficient Adaptation in Multi-Task Learning","author":"stickland","year":"2019","journal-title":"PMLR"},{"key":"ref19","article-title":"BERT-A: Fine-Tuning BERT with Adapters and Data Augmentation","author":"semnani","year":"0"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP40776.2020.9054698"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682775"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462682"},{"key":"ref5","article-title":"On Using Monolingual Corpora in Neural Machine Translation","author":"gulcehre","year":"2015"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"key":"ref7","article-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","author":"devlin","year":"2018"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462506"},{"key":"ref9","article-title":"Language models are unsupervised multitask learners","author":"radford","year":"0"},{"key":"ref1","article-title":"Attention Is All You Need","author":"vaswani","year":"2017","journal-title":"NIPS"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2019-2858"},{"key":"ref22","article-title":"Layer Normalization","author":"ba","year":"2016"},{"key":"ref21","article-title":"Google&#x2019;s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation","author":"wu","year":"2016"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462105"},{"key":"ref23","article-title":"Tensor2Tensor for Neural Machine Translation","author":"vaswani","year":"2018"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2019-2225"}],"event":{"name":"ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","location":"Toronto, ON, Canada","start":{"date-parts":[[2021,6,6]]},"end":{"date-parts":[[2021,6,11]]}},"container-title":["ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9413349\/9413350\/09413475.pdf?arnumber=9413475","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:41:05Z","timestamp":1652197265000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9413475\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,6]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/icassp39728.2021.9413475","relation":{},"subject":[],"published":{"date-parts":[[2021,6,6]]}}}