{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T08:46:13Z","timestamp":1766047573302,"version":"3.33.0"},"reference-count":39,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,12,2]]},"DOI":"10.1109\/slt61566.2024.10832263","type":"proceedings-article","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T18:31:27Z","timestamp":1737052287000},"page":"578-585","source":"Crossref","is-referenced-by-count":1,"title":["Optimizing Dysarthria Wake-Up Word Spotting: an End-to-End Approach For SLT 2024 LRDWWS Challenge"],"prefix":"10.1109","author":[{"given":"Shuiyun","family":"Liu","sequence":"first","affiliation":[{"name":"Northwestern Polytechnical University,Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science,Xi&#x2019;an,China"}]},{"given":"Yuxiang","family":"Kong","sequence":"additional","affiliation":[{"name":"Xiaomi Inc.,China"}]},{"given":"Pengcheng","family":"Guo","sequence":"additional","affiliation":[{"name":"Northwestern Polytechnical University,Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science,Xi&#x2019;an,China"}]},{"given":"Weiji","family":"Zhuang","sequence":"additional","affiliation":[{"name":"Xiaomi Inc.,China"}]},{"given":"Peng","family":"Gao","sequence":"additional","affiliation":[{"name":"Xiaomi Inc.,China"}]},{"given":"Yujun","family":"Wang","sequence":"additional","affiliation":[{"name":"Xiaomi Inc.,China"}]},{"given":"Lei","family":"Xie","sequence":"additional","affiliation":[{"name":"Northwestern Polytechnical University,Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science,Xi&#x2019;an,China"}]}],"member":"263","reference":[{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1109\/ICASSP40776.2020.9053313"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.1109\/ICASSP49357.2023.10096075"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/ICASSP48485.2024.10448040"},{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.21437\/Interspeech.2023-1242"},{"doi-asserted-by":"publisher","key":"ref5","DOI":"10.21437\/Interspeech.2023-1921"},{"doi-asserted-by":"publisher","key":"ref6","DOI":"10.1007\/s40860-019-00085-y"},{"doi-asserted-by":"publisher","key":"ref7","DOI":"10.1109\/icassp43922.2022.9746855"},{"key":"ref8","first-page":"1","article-title":"Improved end-to-end dysarthric speech recognition via meta-learning based model reinitialization","volume-title":"Proc. ISCSLP. 2021","author":"Wang"},{"doi-asserted-by":"publisher","key":"ref9","DOI":"10.1109\/ICASSP49357.2023.10097275"},{"doi-asserted-by":"publisher","key":"ref10","DOI":"10.1109\/TASLP.2021.3091805"},{"doi-asserted-by":"publisher","key":"ref11","DOI":"10.5220\/0009163000002513"},{"doi-asserted-by":"publisher","key":"ref12","DOI":"10.21437\/Interspeech.2019-1206"},{"doi-asserted-by":"publisher","key":"ref13","DOI":"10.21437\/Interspeech.2021-1384"},{"doi-asserted-by":"publisher","key":"ref14","DOI":"10.21437\/Interspeech.2019-1427"},{"doi-asserted-by":"publisher","key":"ref15","DOI":"10.1109\/ICASSP39728.2021.9414644"},{"doi-asserted-by":"publisher","key":"ref16","DOI":"10.21437\/Interspeech.2015-149"},{"key":"ref17","article-title":"Accurate synthesis of dysarthric speech for ASR data augmentation","volume-title":"CoRR","volume":"abs\/2308.08438","author":"Soleymanpour"},{"key":"ref18","article-title":"Synthesizing dysarthric speech using multi-talker TTS for dysarthric speech recognition","volume-title":"CoRR","volume":"abs\/2201.11571","author":"Soleymanpour"},{"volume-title":"Proc. NIPS","author":"Baevski","article-title":"wav2vec 2.0: A framework for self-supervised learning of speech representations","key":"ref19"},{"doi-asserted-by":"publisher","key":"ref20","DOI":"10.21437\/Interspeech.2022-10674"},{"doi-asserted-by":"publisher","key":"ref21","DOI":"10.21437\/Interspeech.2024-879"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.21437\/Interspeech.2022-9996"},{"doi-asserted-by":"publisher","key":"ref23","DOI":"10.21437\/Interspeech.2020-3015"},{"doi-asserted-by":"publisher","key":"ref24","DOI":"10.48550\/ARXIV.1706.03762"},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.18653\/v1\/N19-4009"},{"doi-asserted-by":"publisher","key":"ref26","DOI":"10.1109\/ICASSP40776.2020.9053009"},{"doi-asserted-by":"publisher","key":"ref27","DOI":"10.1145\/1143844.1143891"},{"key":"ref28","article-title":"MUSAN: A music, speech, and noise corpus","volume-title":"CoRR","volume":"abs\/1510.08484","author":"Snyder"},{"doi-asserted-by":"publisher","key":"ref29","DOI":"10.1109\/ICEPE50861.2021.9404373"},{"key":"ref30","first-page":"5530","article-title":"Conditional variational autoencoder with adversarial learning for end-to-end text-to-speech","volume-title":"Proc. ICML. 2021, Proceedings of Machine Learning Research","author":"Kim"},{"key":"ref31","article-title":"Libriheavy: a 50, 000 hours ASR corpus with punctuation casing and context","volume-title":"CoRR","volume":"abs\/2309.08105","author":"Kang"},{"doi-asserted-by":"publisher","key":"ref32","DOI":"10.21437\/Interspeech.2021-1965"},{"doi-asserted-by":"publisher","key":"ref33","DOI":"10.1109\/ICASSP43922.2022.9746682"},{"doi-asserted-by":"publisher","key":"ref34","DOI":"10.1109\/ICSDA.2017.8384449"},{"key":"ref35","article-title":"AISHELL-2: transforming mandarin ASR research into industrial scale","volume-title":"CoRR","volume":"abs\/1808.10583","author":"Du"},{"key":"ref36","first-page":"10254","article-title":"ACAV 100M: automatic curation of large-scale datasets for audio-visual video representation learning","volume-title":"Proc. ICCV. 2021","author":"Lee"},{"doi-asserted-by":"publisher","key":"ref37","DOI":"10.1109\/ICASSP49357.2023.10096736"},{"doi-asserted-by":"publisher","key":"ref38","DOI":"10.1109\/ICASSP.2019.8683474"},{"doi-asserted-by":"publisher","key":"ref39","DOI":"10.1109\/ICASSP49357.2023.10095459"}],"event":{"name":"2024 IEEE Spoken Language Technology Workshop (SLT)","start":{"date-parts":[[2024,12,2]]},"location":"Macao","end":{"date-parts":[[2024,12,5]]}},"container-title":["2024 IEEE Spoken Language Technology Workshop (SLT)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10830790\/10830793\/10832263.pdf?arnumber=10832263","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T08:19:03Z","timestamp":1737101943000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10832263\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,2]]},"references-count":39,"URL":"https:\/\/doi.org\/10.1109\/slt61566.2024.10832263","relation":{},"subject":[],"published":{"date-parts":[[2024,12,2]]}}}