{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,25]],"date-time":"2026-01-25T03:47:43Z","timestamp":1769312863576,"version":"3.49.0"},"reference-count":39,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,5,1]],"date-time":"2019-05-01T00:00:00Z","timestamp":1556668800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,5,1]],"date-time":"2019-05-01T00:00:00Z","timestamp":1556668800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,5,1]],"date-time":"2019-05-01T00:00:00Z","timestamp":1556668800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,5]]},"DOI":"10.1109\/icassp.2019.8683486","type":"proceedings-article","created":{"date-parts":[[2019,4,17]],"date-time":"2019-04-17T20:01:56Z","timestamp":1555531316000},"page":"6740-6744","source":"Crossref","is-referenced-by-count":7,"title":["Attentive Adversarial Learning for Domain-invariant Training"],"prefix":"10.1109","author":[{"given":"Zhong","family":"Meng","sequence":"first","affiliation":[]},{"given":"Jinyu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yifan","family":"Gong","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"131","article-title":"Improving wide-band speech recognition using mixed-bandwidth training data in CD-DNN-HMM","author":"li","year":"2012","journal-title":"Proc SLT"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU.2015.7404837"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1166"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref31","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"radford","year":"2015"},{"key":"ref30","doi-asserted-by":"crossref","DOI":"10.21437\/Interspeech.2017-63","article-title":"Voice conversion from unaligned corpora using variational autoencoding wasserstein generative adversarial networks","author":"hsu","year":"2017"},{"key":"ref37","first-page":"1","article-title":"Multi-channel speech recognition: Lstms all the way through","author":"erdogan","year":"2016","journal-title":"CHiME Workshop"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952160"},{"key":"ref35","article-title":"Long short-term memory recurrent neural network architectures for large scale acoustic modeling","author":"beaufays","year":"2014","journal-title":"InterSpeech"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462497"},{"key":"ref10","article-title":"Unsupervised domain adaptation by backpropagation","author":"ganin","year":"2015","journal-title":"Proc ICML"},{"key":"ref11","first-page":"343","article-title":"Domain separation networks","author":"bousmalis","year":"2016","journal-title":"Proc NIPS"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8461932"},{"key":"ref13","doi-asserted-by":"crossref","DOI":"10.21437\/Interspeech.2017-405","article-title":"English conversational telephone speech recognition by humans and machines","author":"saon","year":"2017"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8461771"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462663"},{"key":"ref16","first-page":"5998","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"Proc NIPS"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1179"},{"key":"ref18","first-page":"577","article-title":"Attention-based models for speech recognition","author":"chorowski","year":"2015","journal-title":"Proc NIPS"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472621"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7953090"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2014.2304637"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-2461"},{"key":"ref3","article-title":"Conversational speech transcription using context-dependent deep neural networks","author":"seide","year":"2011","journal-title":"Proc INTERSPEECH"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2016-879"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683561"},{"key":"ref5","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc NIPS"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8461682"},{"key":"ref7","article-title":"Invariant representations for noisy speech recognition","author":"serdyuk","year":"2016","journal-title":"NIPS Workshop"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/JAS.2017.7510508"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682488"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2012.2205597"},{"key":"ref20","article-title":"Self-attention generative adversarial networks","author":"zhang","year":"2018"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683438"},{"key":"ref21","first-page":"1910","article-title":"Learning small-size DNN with output-distribution-based criteria","author":"li","year":"2014","journal-title":"Proc INTERSPEECH"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682510"},{"key":"ref23","article-title":"Unsupervised adaptation with domain separation networks for robust speech recognition","author":"meng","year":"2017","journal-title":"Proc ASRU"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-2409"},{"key":"ref25","article-title":"Segan: Speech enhancement generative adversarial network","author":"pascual","year":"2017","journal-title":"INTER-SPEECH"}],"event":{"name":"ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","location":"Brighton, United Kingdom","start":{"date-parts":[[2019,5,12]]},"end":{"date-parts":[[2019,5,17]]}},"container-title":["ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8671773\/8682151\/08683486.pdf?arnumber=8683486","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,15]],"date-time":"2022-07-15T03:17:28Z","timestamp":1657855048000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8683486\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,5]]},"references-count":39,"URL":"https:\/\/doi.org\/10.1109\/icassp.2019.8683486","relation":{},"subject":[],"published":{"date-parts":[[2019,5]]}}}