{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,7]],"date-time":"2024-09-07T13:23:17Z","timestamp":1725715397358},"reference-count":27,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,11]]},"DOI":"10.1109\/apsipaasc47483.2019.9023352","type":"proceedings-article","created":{"date-parts":[[2020,3,6]],"date-time":"2020-03-06T12:03:54Z","timestamp":1583496234000},"page":"284-288","source":"Crossref","is-referenced-by-count":1,"title":["Phonetic-Attention Scoring for Deep Speaker Features in Speaker Verification"],"prefix":"10.1109","author":[{"given":"Lantian","family":"Li","sequence":"first","affiliation":[]},{"given":"Zhiyuan","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Dong","family":"Wang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"crossref","first-page":"1542","DOI":"10.21437\/Interspeech.2017-452","article-title":"Deep speaker feature learning for text-independent speaker verification","author":"li","year":"2017","journal-title":"InterSpeech"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/APSIPA.2017.8282083"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472652"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2016.7846261"},{"key":"ref14","doi-asserted-by":"crossref","DOI":"10.21437\/Interspeech.2017-1608","article-title":"End-to-end text-independent speaker verification with triplet loss on short utterances","author":"zhang","year":"2017","journal-title":"InterSpeech"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/APSIPA.2017.8282024"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2018.8462169"},{"key":"ref17","article-title":"Neural machine translation by jointly learning to align and translate","author":"bahdanau","year":"2015","journal-title":"ICLRE"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472621"},{"key":"ref19","first-page":"5359","article-title":"Attention-based models for text-dependent speaker verification","author":"chowdhury","year":"2018","journal-title":"ICASSP"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/11744085_41"},{"key":"ref27","article-title":"The kaldi speech recognition toolkit","author":"povey","year":"0","journal-title":"Workshop on Automatic Speech Recognition and Understanding"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2064307"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2014.6854363"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.21236\/ADA613971"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2016.7846260"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2015.07.003"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2006.881693"},{"key":"ref9","doi-asserted-by":"crossref","first-page":"999","DOI":"10.21437\/Interspeech.2017-620","article-title":"Deep neural network embeddings for text-independent speaker verification","author":"snyder","year":"2017","journal-title":"InterSpeech"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1006\/dspr.1999.0361"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ISCSLP.2018.8706589"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-993"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-1158"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2018.8639574"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-1226"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7953152"},{"key":"ref25","article-title":"MUSAN: A Music, Speech, and Noise Corpus","author":"snyder","year":"2015","journal-title":"ArXiv"}],"event":{"name":"2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","start":{"date-parts":[[2019,11,18]]},"location":"Lanzhou, China","end":{"date-parts":[[2019,11,21]]}},"container-title":["2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8989870\/9023008\/09023352.pdf?arnumber=9023352","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,17]],"date-time":"2022-07-17T17:55:00Z","timestamp":1658080500000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9023352\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,11]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/apsipaasc47483.2019.9023352","relation":{},"subject":[],"published":{"date-parts":[[2019,11]]}}}