{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,17]],"date-time":"2026-04-17T03:19:27Z","timestamp":1776395967909,"version":"3.51.2"},"reference-count":98,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2021,10,18]],"date-time":"2021-10-18T00:00:00Z","timestamp":1634515200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2021,10,18]],"date-time":"2021-10-18T00:00:00Z","timestamp":1634515200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Nat Mach Intell"],"DOI":"10.1038\/s42256-021-00394-z","type":"journal-article","created":{"date-parts":[[2021,10,19]],"date-time":"2021-10-19T00:13:50Z","timestamp":1634602430000},"page":"840-849","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":55,"title":["Harnessing the power of artificial intelligence to transform hearing healthcare and research"],"prefix":"10.1038","volume":"3","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5238-4462","authenticated-orcid":false,"given":"Nicholas A.","family":"Lesica","sequence":"first","affiliation":[]},{"given":"Nishchay","family":"Mehta","sequence":"additional","affiliation":[]},{"given":"Joseph G.","family":"Manjaly","sequence":"additional","affiliation":[]},{"given":"Li","family":"Deng","sequence":"additional","affiliation":[]},{"given":"Blake S.","family":"Wilson","sequence":"additional","affiliation":[]},{"given":"Fan-Gang","family":"Zeng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,10,18]]},"reference":[{"key":"394_CR1","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.heares.2008.06.005","volume":"242","author":"BS Wilson","year":"2008","unstructured":"Wilson, B. S. & Dorman, M. F. Cochlear implants: a remarkable past and a brilliant future. Hear. Res. 242, 3\u201321 (2008).","journal-title":"Hear. Res."},{"key":"394_CR2","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1109\/RBME.2008.2008250","volume":"1","author":"F-G Zeng","year":"2008","unstructured":"Zeng, F.-G., Rebscher, S., Harrison, W. V., Sun, X. & Feng, H. Cochlear implants: system design, integration and evaluation. IEEE Rev. Biomed. Eng. 1, 115\u2013142 (2008).","journal-title":"IEEE Rev. Biomed. Eng."},{"key":"394_CR3","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1177\/1084713806298000","volume":"11","author":"H Levitt","year":"2007","unstructured":"Levitt, H. A historical perspective on digital hearing aids: how digital technology has changed modern hearing aids. Trends Amplif. 11, 7\u201324 (2007).","journal-title":"Trends Amplif."},{"key":"394_CR4","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MSP.2012.2205597","volume":"29","author":"G Hinton","year":"2012","unstructured":"Hinton, G. et al. Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process. Mag. 29, 82\u201397 (2012).","journal-title":"IEEE Signal Process. Mag."},{"key":"394_CR5","doi-asserted-by":"crossref","unstructured":"Yu, D. & Deng, L. Automatic Speech Recognition - A Deep Learning Approach (Springer, 2015).","DOI":"10.1007\/978-1-4471-5779-3"},{"key":"394_CR6","doi-asserted-by":"crossref","unstructured":"Deng, L. & Liu, Y. Deep Learning in Natural Language Processing (Springer, 2018).","DOI":"10.1007\/978-981-10-5209-5"},{"key":"394_CR7","unstructured":"Devlin, J., Chang, M. W., Lee, K. & Toutanova, K. BERT: pre-training of deep bidirectional transformers for language understanding. Preprint at https:\/\/arxiv.org\/abs\/1810.04805 (2018)."},{"key":"394_CR8","doi-asserted-by":"publisher","unstructured":"Lindsay, G.W. Convolutional neural networks as a model of the visual system: past, present, and future. J. Cogn. Neurosci. https:\/\/doi.org\/10.1162\/jocn_a_01544 (2020).","DOI":"10.1162\/jocn_a_01544"},{"key":"394_CR9","doi-asserted-by":"publisher","first-page":"176","DOI":"10.1016\/j.conb.2020.11.009","volume":"65","author":"RS van Bergen","year":"2020","unstructured":"van Bergen, R. S. & Kriegeskorte, N. Going in circles is the way forward: the role of recurrence in visual inference. Curr. Opin. Neurobiol. 65, 176\u2013193 (2020).","journal-title":"Curr. Opin. Neurobiol."},{"key":"394_CR10","unstructured":"Lohr, S. Frederick Jelinek, who gave machines the key to human speech, dies at 77. The New York Times (24 September 2010)."},{"key":"394_CR11","doi-asserted-by":"publisher","unstructured":"Wasmann, J.-W. A. et al. Computational audiology: new approaches to advance hearing health care in the digital age. Ear Hear. https:\/\/doi.org\/10.1097\/AUD.0000000000001041 (2021).","DOI":"10.1097\/AUD.0000000000001041"},{"key":"394_CR12","doi-asserted-by":"publisher","first-page":"145","DOI":"10.1109\/TBCAS.2020.2974154","volume":"14","author":"Y Wei","year":"2020","unstructured":"Wei, Y. et al. A review of algorithm hardware design for AI-based biomedical applications. IEEE Trans. Biomed. Circuits Syst. 14, 145\u2013163 (2020).","journal-title":"IEEE Trans. Biomed. Circuits Syst."},{"key":"394_CR13","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1109\/JSTSP.2019.2908700","volume":"13","author":"H Purwins","year":"2019","unstructured":"Purwins, H. et al. Deep learning for audio signal processing. IEEE J. Sel. Top. Signal Process. 13, 206\u2013219 (2019).","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"394_CR14","unstructured":"World Report on Hearing (World Health Organization, 2021)."},{"key":"394_CR15","doi-asserted-by":"publisher","first-page":"465","DOI":"10.1016\/S0140-6736(04)15495-0","volume":"363","author":"MM Rovers","year":"2004","unstructured":"Rovers, M. M., Schilder, A. G., Zielhuis, G. A. & Rosenfeld, R. M. Otitis media. Lancet 363, 465\u2013473 (2004).","journal-title":"Lancet"},{"key":"394_CR16","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1001\/archpedi.155.10.1137","volume":"155","author":"ME Pichichero","year":"2001","unstructured":"Pichichero, M. E. & Poole, M. D. Assessing diagnostic accuracy and tympanocentesis skills in the management of otitis media. Arch. Pediatr. Adolesc. Med. 155, 1137\u20131142 (2001).","journal-title":"Arch. Pediatr. Adolesc. Med."},{"key":"394_CR17","doi-asserted-by":"publisher","first-page":"606","DOI":"10.1016\/j.ebiom.2019.06.050","volume":"45","author":"D Cha","year":"2019","unstructured":"Cha, D., Pae, C., Seong, S.-B., Choi, J. Y. & Park, H.-J. Automated diagnosis of ear disease using ensemble deep learning with a big otoendoscopy image database. EBioMedicine 45, 606\u2013614 (2019).","journal-title":"EBioMedicine"},{"key":"394_CR18","unstructured":"World\u2019s first otoscope with artificial intelligence (AI) image classification of ear diseases. hearX https:\/\/hearxgroup.com\/blog\/world-first-otoscope-ai-image-classification-of-ear-diseases.html (2020)."},{"key":"394_CR19","unstructured":"Hearing Health (Health Informatics Collaborative, nihr, 2020); https:\/\/hic.nihr.ac.uk\/hearing+health"},{"key":"394_CR20","unstructured":"Sing Registry: The Genetic Sensorineural Hearing Loss Registry (Sing, 2020); http:\/\/singregistry.com"},{"key":"394_CR21","doi-asserted-by":"publisher","first-page":"195","DOI":"10.3109\/14992020903470783","volume":"49","author":"DW Swanepoel","year":"2010","unstructured":"Swanepoel, D. W. et al. Telehealth in audiology: the need and potential to reach underserved communities. Int. J. Audiol. 49, 195\u2013202 (2010).","journal-title":"Int. J. Audiol."},{"key":"394_CR22","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1097\/01.HJ.0000669852.90548.75","volume":"73","author":"DW Swanepoel","year":"2020","unstructured":"Swanepoel, D. W. & Hall, J. W. Making audiology work during COVID-19 and beyond. Hear. j. 73, 20\u201324 (2020).","journal-title":"Hear. j"},{"key":"394_CR23","doi-asserted-by":"publisher","first-page":"918","DOI":"10.1097\/AUD.0000000000000669","volume":"40","author":"DL Barbour","year":"2019","unstructured":"Barbour, D. L. et al. Online machine learning audiometry. Ear Hear. 40, 918\u2013926 (2019).","journal-title":"Ear Hear."},{"key":"394_CR24","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1121\/1.5047436","volume":"144","author":"J Schlittenlacher","year":"2018","unstructured":"Schlittenlacher, J., Turner, R. E. & Moore, B. C. J. Audiogram estimation using Bayesian active learning. J. Acoust. Soc. Am. 144, 421\u2013430 (2018).","journal-title":"J. Acoust. Soc. Am."},{"key":"394_CR25","doi-asserted-by":"publisher","first-page":"578","DOI":"10.1177\/0003489420902162","volume":"129","author":"J Sandstr\u00f6m","year":"2020","unstructured":"Sandstr\u00f6m, J., Swanepoel, D., Laurent, C., Umefjord, G. & Lundberg, T. Accuracy and reliability of smartphone self-test audiometry in community clinics in low income settings: a comparative study. Ann. Otol. Rhinol. Laryngol. 129, 578\u2013584 (2020).","journal-title":"Ann. Otol. Rhinol. Laryngol."},{"key":"394_CR26","doi-asserted-by":"publisher","first-page":"198","DOI":"10.1080\/14670100.2017.1325093","volume":"18","author":"M Meeuws","year":"2017","unstructured":"Meeuws, M. et al. Computer-assisted CI fitting: is the learning capacity of the intelligent agent FOX beneficial for speech understanding? Cochlear Implants Int. 18, 198\u2013206 (2017).","journal-title":"Cochlear Implants Int."},{"key":"394_CR27","doi-asserted-by":"publisher","first-page":"260","DOI":"10.1080\/14670100.2020.1757840","volume":"21","author":"M Meeuws","year":"2020","unstructured":"Meeuws, M., Pascoal, D., Janssens de Varebeke, S., De Ceulaer, G. & Govaerts, P. J. Cochlear implant telemedicine: remote fitting based on psychoacoustic self-tests and artificial intelligence. Cochlear Implants Int. 21, 260\u2013268 (2020).","journal-title":"Cochlear Implants Int."},{"key":"394_CR28","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1016\/j.tins.2018.01.008","volume":"41","author":"NA Lesica","year":"2018","unstructured":"Lesica, N. A. Why do hearing aids fail to restore normal auditory perception? Trends Neurosci. 41, 174\u2013185 (2018).","journal-title":"Trends Neurosci."},{"key":"394_CR29","doi-asserted-by":"publisher","first-page":"24","DOI":"10.1016\/j.heares.2014.11.009","volume":"322","author":"BS Wilson","year":"2015","unstructured":"Wilson, B. S. Getting a decent (but sparse) signal to the brain for users of cochlear implants. Hear. Res. 322, 24\u201338 (2015).","journal-title":"Hear. Res."},{"key":"394_CR30","doi-asserted-by":"publisher","first-page":"1662","DOI":"10.1109\/TBME.2017.2718939","volume":"64","author":"F-G Zeng","year":"2017","unstructured":"Zeng, F.-G. Challenges in improving cochlear implant performance and accessibility. IEEE Trans. Biomed. Eng. 64, 1662\u20131664 (2017).","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"394_CR31","first-page":"6","volume":"68","author":"F-G Zeng","year":"2015","unstructured":"Zeng, F.-G. Do or die for hearing aid industry. Hear. J. 68, 6 (2015).","journal-title":"Hear. J."},{"key":"394_CR32","unstructured":"Oticon: More Technology Polaris For Professionals (Oticon); https:\/\/www.oticon.com\/professionals\/brainhearing-technology\/more-technology"},{"key":"394_CR33","unstructured":"Artificial Intelligence in Hearing Aids (Widex Professionals); https:\/\/uk.widex.pro\/en-gb\/evidence-technology\/technological-excellence\/artificial-intelligence-in-hearing-aids"},{"key":"394_CR34","doi-asserted-by":"publisher","first-page":"131S","DOI":"10.1097\/AUD.0000000000000955","volume":"41","author":"M Slaney","year":"2020","unstructured":"Slaney, M. et al. Auditory measures for the next billion users. Ear Hear. 41, 131S (2020).","journal-title":"Ear Hear."},{"key":"394_CR35","doi-asserted-by":"publisher","first-page":"1702","DOI":"10.1109\/TASLP.2018.2842159","volume":"26","author":"D Wang","year":"2018","unstructured":"Wang, D. & Chen, J. Supervised speech separation based on deep learning: an overview. IEEEACM Trans. Audio Speech Lang. Process. 26, 1702\u20131726 (2018).","journal-title":"IEEEACM Trans. Audio Speech Lang. Process."},{"key":"394_CR36","unstructured":"Whisper: The New Hearing Aid That Gets Better Over Time (Whisper); https:\/\/whisper.ai\/"},{"key":"394_CR37","unstructured":"HD Voice with Echo & Noise Cancellation (Krisp); https:\/\/www.krisp.ai\/"},{"key":"394_CR38","doi-asserted-by":"publisher","first-page":"4106","DOI":"10.1121\/10.0001441","volume":"147","author":"EW Healy","year":"2020","unstructured":"Healy, E. W., Johnson, E. M., Delfarah, M. & Wang, D. A talker-independent deep learning algorithm to increase intelligibility for hearing-impaired listeners in reverberant competing talker conditions. J. Acoust. Soc. Am. 147, 4106 (2020).","journal-title":"J. Acoust. Soc. Am."},{"key":"394_CR39","doi-asserted-by":"publisher","first-page":"705","DOI":"10.1121\/1.5119226","volume":"146","author":"T Goehring","year":"2019","unstructured":"Goehring, T., Keshavarzi, M., Carlyon, R. P. & Moore, B. C. J. Using recurrent neural networks to improve the perception of speech in non-stationary noise by people with cochlear implants. J. Acoust. Soc. Am. 146, 705 (2019).","journal-title":"J. Acoust. Soc. Am."},{"key":"394_CR40","doi-asserted-by":"publisher","first-page":"795","DOI":"10.1097\/AUD.0000000000000537","volume":"39","author":"Y-H Lai","year":"2018","unstructured":"Lai, Y.-H. et al. Deep learning\u2013based noise reduction approach to improve speech intelligibility for cochlear implant recipients. Ear Hear. 39, 795\u2013809 (2018).","journal-title":"Ear Hear."},{"key":"394_CR41","doi-asserted-by":"publisher","first-page":"184","DOI":"10.1109\/TNSRE.2020.3042655","volume":"29","author":"NY-H Wang","year":"2021","unstructured":"Wang, N. Y.-H. et al. Improving the intelligibility of speech for simulated electric and acoustic stimulation using fully convolutional neural networks. IEEE Trans. Neural Syst. Rehabil. Eng. 29, 184\u2013195 (2021).","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"key":"394_CR42","doi-asserted-by":"publisher","unstructured":"An, W. W., Pei, A., Noyce, A. L. & Shinn-Cunningham, B. Decoding auditory attention from single-trial EEG for a high-efficiency brain-computer interface. In 2020 42nd Annual International Conference of the IEEE Engineering in Medicine Biology Society (EMBC) 3456\u20133459 (IEEE, 2020); https:\/\/doi.org\/10.1109\/EMBC44109.2020.9175753","DOI":"10.1109\/EMBC44109.2020.9175753"},{"key":"394_CR43","doi-asserted-by":"publisher","first-page":"036020","DOI":"10.1088\/1741-2552\/aa66dd","volume":"14","author":"L Fiedler","year":"2017","unstructured":"Fiedler, L. et al. Single-channel in-ear-EEG detects the focus of auditory attention to concurrent tone streams and mixed speech. J. Neural Eng. 14, 036020 (2017).","journal-title":"J. Neural Eng."},{"key":"394_CR44","doi-asserted-by":"crossref","unstructured":"O\u2019Sullivan, J. A. et al. Attentional selection in a cocktail party environment can be decoded from single-trial EEG. Cereb. Cortex 25, 1697\u20131706 (2015).","DOI":"10.1093\/cercor\/bht355"},{"key":"394_CR45","doi-asserted-by":"publisher","first-page":"140S","DOI":"10.1097\/AUD.0000000000000961","volume":"41","author":"R Mehra","year":"2020","unstructured":"Mehra, R., Brimijoin, O., Robinson, P. & Lunner, T. Potential of augmented reality platforms to improve individual hearing aids and to support more ecologically valid research. Ear Hear. 41, 140S\u2013146S (2020).","journal-title":"Ear Hear."},{"key":"394_CR46","doi-asserted-by":"publisher","unstructured":"Tseng, R.-Y. et al. A study of joint effect on denoising techniques and visual cues to improve speech intelligibility in cochlear implant simulation. IEEE Trans. Cogn. Dev. Syst. https:\/\/doi.org\/10.1109\/TCDS.2020.3017042 (2020).","DOI":"10.1109\/TCDS.2020.3017042"},{"key":"394_CR47","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1080\/14992027.2017.1367848","volume":"57","author":"T Goehring","year":"2018","unstructured":"Goehring, T., Chapman, J. L., Bleeck, S. & Monaghan, J. J. M. Tolerable delay for speech production and perception: effects of hearing ability and experience with hearing aids. Int. J. Audiol. 57, 61\u201368 (2018).","journal-title":"Int. J. Audiol."},{"key":"394_CR48","doi-asserted-by":"publisher","first-page":"1600","DOI":"10.1016\/S0140-6736(13)60142-7","volume":"382","author":"D Baguley","year":"2013","unstructured":"Baguley, D., McFerran, D. & Hall, D. Tinnitus. Lancet 382, 1600\u20131607 (2013).","journal-title":"Lancet"},{"key":"394_CR49","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.neuron.2019.05.008","volume":"103","author":"SE Shore","year":"2019","unstructured":"Shore, S. E. & Wu, C. Mechanisms of noise-induced tinnitus: insights from cellular studies. Neuron 103, 8\u201320 (2019).","journal-title":"Neuron"},{"key":"394_CR50","doi-asserted-by":"publisher","first-page":"732","DOI":"10.3389\/fneur.2017.00732","volume":"8","author":"V Iliadou","year":"2018","unstructured":"Iliadou, V. & Kiese-Himmel, C. Common misconceptions regarding pediatric auditory processing disorder. Front. Neurol. 8, 732 (2018).","journal-title":"Front. Neurol."},{"key":"394_CR51","doi-asserted-by":"publisher","first-page":"287","DOI":"10.3389\/fneur.2019.00287","volume":"10","author":"K Neijenhuis","year":"2019","unstructured":"Neijenhuis, K. et al. An evidence-based perspective on \u2018misconceptions\u2019 regarding pediatric auditory processing disorder. Front. Neurol. 10, 287 (2019).","journal-title":"Front. Neurol."},{"key":"394_CR52","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-018-02820-0","volume":"9","author":"S Vanneste","year":"2018","unstructured":"Vanneste, S., Song, J.-J. & De Ridder, D. Thalamocortical dysrhythmia detected by machine learning. Nat. Commun. 9, 1103 (2018).","journal-title":"Nat. Commun."},{"key":"394_CR53","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1159\/000504741","volume":"25","author":"G Palacios","year":"2020","unstructured":"Palacios, G., Nore\u00f1a, A. & Londero, A. Assessing the heterogeneity of complaints related to tinnitus and hyperacusis from an unsupervised machine learning approach: an exploratory study. Audiol. Neurootol. 25, 174\u2013189 (2020).","journal-title":"Audiol. Neurootol."},{"key":"394_CR54","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1016\/j.heares.2017.12.018","volume":"360","author":"S Verhulst","year":"2018","unstructured":"Verhulst, S., Alto\u00e8, A. & Vasilkov, V. Computational modeling of the human auditory periphery: auditory-nerve responses, evoked potentials and hearing loss. Hear. Res. 360, 55\u201375 (2018).","journal-title":"Hear. Res."},{"key":"394_CR55","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1016\/j.heares.2015.06.014","volume":"330","author":"C Kopp-Scheinpflug","year":"2015","unstructured":"Kopp-Scheinpflug, C. & Tempel, B. L. Decreased temporal precision of neuronal signaling as a candidate mechanism of auditory processing disorder. Hear. Res. 330, 213\u2013220 (2015).","journal-title":"Hear. Res."},{"key":"394_CR56","doi-asserted-by":"publisher","first-page":"460","DOI":"10.3758\/BF03210983","volume":"2","author":"ME Farmer","year":"1995","unstructured":"Farmer, M. E. & Klein, R. M. The evidence for a temporal processing deficit linked to dyslexia: a review. Psychon. Bull. Rev. 2, 460\u2013493 (1995).","journal-title":"Psychon. Bull. Rev."},{"key":"394_CR57","doi-asserted-by":"publisher","first-page":"150","DOI":"10.1016\/j.bandc.2007.12.005","volume":"67","author":"CA Carroll","year":"2008","unstructured":"Carroll, C. A., Boggs, J., O\u2019Donnell, B. F., Shekhar, A. & Hetrick, W. P. Temporal processing dysfunction in schizophrenia. Brain Cogn. 67, 150\u2013161 (2008).","journal-title":"Brain Cogn."},{"key":"394_CR58","doi-asserted-by":"publisher","first-page":"983","DOI":"10.1152\/physrev.00026.2009","volume":"90","author":"B Grothe","year":"2010","unstructured":"Grothe, B., Pecka, M. & McAlpine, D. Mechanisms of sound localization in mammals. Physiol. Rev. 90, 983\u20131012 (2010).","journal-title":"Physiol. Rev."},{"key":"394_CR59","doi-asserted-by":"publisher","first-page":"687","DOI":"10.1016\/j.neuron.2018.03.045","volume":"98","author":"JJ Paton","year":"2018","unstructured":"Paton, J. J. & Buonomano, D. V. The neural basis of timing: distributed mechanisms for diverse functions. Neuron 98, 687\u2013705 (2018).","journal-title":"Neuron"},{"key":"394_CR60","doi-asserted-by":"publisher","first-page":"e11386","DOI":"10.7554\/eLife.11386","volume":"4","author":"TS Gouv\u00eaa","year":"2015","unstructured":"Gouv\u00eaa, T. S. et al. Striatal dynamics explain duration judgments. eLife 4, e11386 (2015).","journal-title":"eLife"},{"key":"394_CR61","unstructured":"van den Oord, A. et al. WaveNet: a generative model for raw audio. Preprint at https:\/\/arxiv.org\/abs\/1609.03499 (2016)."},{"key":"394_CR62","unstructured":"Vaswani, A. et al. Attention is all you need. Preprint at https:\/\/arxiv.org\/abs\/1706.03762 (2017)."},{"key":"394_CR63","doi-asserted-by":"publisher","first-page":"10530","DOI":"10.1073\/pnas.1921609117","volume":"117","author":"Z Bi","year":"2020","unstructured":"Bi, Z. & Zhou, C. Understanding the computation of time using neural network models. Proc. Natl Acad. Sci. USA 117, 10530\u201310540 (2020).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"394_CR64","doi-asserted-by":"publisher","first-page":"394","DOI":"10.1073\/pnas.1619449114","volume":"114","author":"JD Murray","year":"2017","unstructured":"Murray, J. D. et al. Stable population coding for working memory coexists with heterogeneous neural dynamics in prefrontal cortex. Proc. Natl Acad. Sci. USA 114, 394\u2013399 (2017).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"394_CR65","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1038\/s41593-017-0028-6","volume":"21","author":"J Wang","year":"2018","unstructured":"Wang, J., Narain, D., Hosseini, E. A. & Jazayeri, M. Flexible timing by temporal scaling of cortical responses. Nat. Neurosci. 21, 102\u2013110 (2018).","journal-title":"Nat. Neurosci."},{"key":"394_CR66","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1109\/JSTSP.2020.2987728","volume":"14","author":"C Zhang","year":"2020","unstructured":"Zhang, C., Yang, Z., He, X. & Deng, L. Multimodal intelligence: representation learning, information fusion, and applications. IEEE J. Sel. Top. Signal Process. 14, 478\u2013493 (2020).","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"394_CR67","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.cophys.2020.09.003","volume":"18","author":"JK Bizley","year":"2020","unstructured":"Bizley, J. K. & Dai, Y. Non-auditory processing in the central auditory pathway. Curr. Opin. Physiol. 18, 100\u2013105 (2020).","journal-title":"Curr. Opin. Physiol."},{"key":"394_CR68","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1038\/s41593-018-0310-2","volume":"22","author":"GR Yang","year":"2019","unstructured":"Yang, G. R., Joglekar, M. R., Song, H. F., Newsome, W. T. & Wang, X.-J. Task representations in neural networks trained to perform many cognitive tasks. Nat. Neurosci. 22, 297\u2013306 (2019).","journal-title":"Nat. Neurosci."},{"key":"394_CR69","unstructured":"Brown, T. B. et al. Language models are few-shot learners. Preprint at https:\/\/arxiv.org\/abs\/2005.14165 (2020)."},{"key":"394_CR70","doi-asserted-by":"publisher","unstructured":"Deng, L., Hinton, G. & Kingsbury, B. New types of deep neural network learning for speech recognition and related applications: an overview. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing 8599\u20138603 (IEEE, 2013); https:\/\/doi.org\/10.1109\/ICASSP.2013.6639344","DOI":"10.1109\/ICASSP.2013.6639344"},{"key":"394_CR71","doi-asserted-by":"publisher","first-page":"6007","DOI":"10.1523\/JNEUROSCI.0396-20.2020","volume":"40","author":"F-G Zeng","year":"2020","unstructured":"Zeng, F.-G., Richardson, M. & Turner, K. Tinnitus does not interfere with auditory and speech perception. J. Neurosci. 40, 6007\u20136017 (2020).","journal-title":"J. Neurosci."},{"key":"394_CR72","doi-asserted-by":"publisher","first-page":"123","DOI":"10.1016\/j.cophys.2020.10.009","volume":"18","author":"F-G Zeng","year":"2020","unstructured":"Zeng, F.-G. Tinnitus and hyperacusis: central noise, gain and variance. Curr. Opin. Physiol. 18, 123\u2013129 (2020).","journal-title":"Curr. Opin. Physiol."},{"key":"394_CR73","doi-asserted-by":"publisher","first-page":"150","DOI":"10.1038\/nrneurol.2016.12","volume":"12","author":"SE Shore","year":"2016","unstructured":"Shore, S. E., Roberts, L. E. & Langguth, B. Maladaptive plasticity in tinnitus\u2013triggers, mechanisms and treatment. Nat. Rev. Neurol. 12, 150\u2013160 (2016).","journal-title":"Nat. Rev. Neurol."},{"key":"394_CR74","doi-asserted-by":"publisher","first-page":"613","DOI":"10.1001\/jamaoto.2020.0662","volume":"146","author":"EE Zhao","year":"2020","unstructured":"Zhao, E. E. et al. Association of patient-related factors with adult cochlear implant speech recognition outcomes: a meta-analysis. JAMA Otolaryngol. Head Neck Surg. 146, 613\u2013620 (2020).","journal-title":"JAMA Otolaryngol. Head Neck Surg."},{"key":"394_CR75","doi-asserted-by":"publisher","first-page":"e12394","DOI":"10.1111\/lnc3.12394","volume":"14","author":"JR Drouin","year":"2020","unstructured":"Drouin, J. R. & Theodore, R. M. Leveraging interdisciplinary perspectives to optimize auditory training for cochlear implant users. Lang. Linguist. Compass 14, e12394 (2020).","journal-title":"Lang. Linguist. Compass"},{"key":"394_CR76","doi-asserted-by":"publisher","first-page":"630","DOI":"10.1016\/j.neuron.2018.03.044","volume":"98","author":"AJE Kell","year":"2018","unstructured":"Kell, A. J. E., Yamins, D. L. K., Shook, E. N., Norman-Haignere, S. V. & McDermott, J. H. A task-optimized neural network replicates human auditory behavior, predicts brain responses, and reveals a cortical processing hierarchy. Neuron 98, 630\u2013644 (2018).","journal-title":"Neuron"},{"key":"394_CR77","doi-asserted-by":"publisher","first-page":"e53445","DOI":"10.7554\/eLife.53445","volume":"9","author":"M Keshishian","year":"2020","unstructured":"Keshishian, M. et al. Estimating and interpreting nonlinear receptive field of sensory neural responses with deep neural network models. eLife 9, e53445 (2020).","journal-title":"eLife"},{"key":"394_CR78","doi-asserted-by":"publisher","unstructured":"Baby, D., Van Den Broucke, A. & Verhulst, S. A convolutional neural-network model of human cochlear mechanics and filter tuning for real-time applications. Nat. Mach. Intell. https:\/\/doi.org\/10.1038\/s42256-020-00286-8 (2021).","DOI":"10.1038\/s42256-020-00286-8"},{"key":"394_CR79","doi-asserted-by":"publisher","first-page":"1060","DOI":"10.1109\/TASL.2013.2244083","volume":"21","author":"L Deng","year":"2013","unstructured":"Deng, L. & Li, X. Machine learning paradigms for speech recognition: An overview. IEEE Trans. Audio Speech Lang. Process. 21, 1060\u20131089 (2013).","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"394_CR80","doi-asserted-by":"publisher","unstructured":"Deng, L., Abdel-Hamid, O. & Yu, D. A deep convolutional neural network using heterogeneous pooling for trading acoustic invariance with phonetic confusion. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing 6669\u20136673 (IEEE, 2013); https:\/\/doi.org\/10.1109\/ICASSP.2013.6638952","DOI":"10.1109\/ICASSP.2013.6638952"},{"key":"394_CR81","doi-asserted-by":"publisher","unstructured":"Schofield, B. R. in Auditory and Vestibular Efferents (eds Ryugo, D. K. & Fay, R. R.) 261\u2013290 (Springer, 2011); https:\/\/doi.org\/10.1007\/978-1-4419-7070-1_9","DOI":"10.1007\/978-1-4419-7070-1_9"},{"key":"394_CR82","unstructured":"Sabour, S., Frosst, N. & Hinton, G. E. Dynamic routing between capsules. In Advances in Neural Information Processing Systems 3856\u20133866 (NIPS, 2017)."},{"key":"394_CR83","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1016\/j.visres.2019.12.006","volume":"167","author":"A Doerig","year":"2020","unstructured":"Doerig, A., Bornet, A., Choung, O. H. & Herzog, M. H. Crowding reveals fundamental differences in local vs. global processing in humans and machines. Vision Res. 167, 39\u201345 (2020).","journal-title":"Vision Res."},{"key":"394_CR84","doi-asserted-by":"publisher","unstructured":"Zhang, M., Denison, R. N., Pelli, D. G., Le, T. T. C. & Ihlefeld, A. Informational masking vs. crowding\u2014a mid-level trade-off between auditory and visual processing. Preprint at bioRxiv https:\/\/doi.org\/10.1101\/2021.04.21.440826 (2021).","DOI":"10.1101\/2021.04.21.440826"},{"key":"394_CR85","unstructured":"Hinton, G. How to represent part-whole hierarchies in a neural network. Preprint at https:\/\/arxiv.org\/abs\/2102.12627 (2021)."},{"key":"394_CR86","unstructured":"Committee on Accessible and Affordable Hearing Health Care for Adults, Board on Health Sciences Policy, Health and Medicine Division & National Academies of Sciences, Engineering, and Medicine Hearing Health Care for Adults: Priorities for Improving Access and Affordability (National Academies, 2016)."},{"key":"394_CR87","unstructured":"Aging America & Hearing Loss: Imperative of Improved Hearing Technologies (President\u2019s Council of Advisors on Science and Technology, 2016)."},{"key":"394_CR88","doi-asserted-by":"publisher","first-page":"609","DOI":"10.1001\/jamainternmed.2017.0464","volume":"177","author":"E Warren","year":"2017","unstructured":"Warren, E. & Grassley, C. Over-the-counter hearing aids: the path forward. JAMA Intern. Med. 177, 609\u2013610 (2017).","journal-title":"JAMA Intern. Med."},{"key":"394_CR89","unstructured":"Davies-Venn, E. & Glista, D. Connected hearing healthcare: the realisation of benefit relies on successful clinical implementation. ENT & Audiology News https:\/\/www.entandaudiologynews.com\/features\/audiology-features\/post\/connected-hearing-healthcare-the-realisation-of-benefit-relies-on-successful-clinical-implementation (2019)."},{"key":"394_CR90","doi-asserted-by":"publisher","first-page":"2141","DOI":"10.1001\/jama.2020.5035","volume":"323","author":"CJ Lindsell","year":"2020","unstructured":"Lindsell, C. J., Stead, W. W. & Johnson, K. B. Action-informed artificial intelligence\u2014matching the algorithm to the problem. JAMA 323, 2141\u20132142 (2020).","journal-title":"JAMA"},{"key":"394_CR91","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1109\/MSP.2010.937498","volume":"27","author":"RF Lyon","year":"2010","unstructured":"Lyon, R. F. Machine hearing: an emerging field. IEEE Signal Process Mag. 27, 131\u2013139 (2010).","journal-title":"IEEE Signal Process Mag."},{"key":"394_CR92","doi-asserted-by":"publisher","first-page":"132","DOI":"10.1080\/14992027.2018.1543958","volume":"58","author":"S Denys","year":"2019","unstructured":"Denys, S., Latzel, M., Francart, T. & Wouters, J. A preliminary investigation into hearing aid fitting based on automated real-ear measurements integrated in the fitting software: test\u2013retest reliability, matching accuracy and perceptual outcomes. Int. J. Audiol. 58, 132\u2013140 (2019).","journal-title":"Int. J. Audiol."},{"key":"394_CR93","doi-asserted-by":"publisher","first-page":"E1022","DOI":"10.1073\/pnas.1717603115","volume":"115","author":"G Feng","year":"2018","unstructured":"Feng, G. et al. Neural preservation underlies speech improvement from auditory deprivation in young cochlear implant recipients. Proc. Natl Acad. Sci. USA 115, E1022\u2013E1031 (2018).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"394_CR94","doi-asserted-by":"publisher","unstructured":"Zhou, Z. et al. Sign-to-speech translation using machine-learning-assisted stretchable sensor arrays. Nat. Electron. https:\/\/doi.org\/10.1038\/s41928-020-0428-6 (2020).","DOI":"10.1038\/s41928-020-0428-6"},{"key":"394_CR95","doi-asserted-by":"publisher","first-page":"1618","DOI":"10.1121\/1.4960486","volume":"140","author":"A Saremi","year":"2016","unstructured":"Saremi, A. et al. A comparative study of seven human cochlear filter models. J. Acoust. Soc. Am. 140, 1618\u20131634 (2016).","journal-title":"J. Acoust. Soc. Am."},{"key":"394_CR96","doi-asserted-by":"publisher","first-page":"925","DOI":"10.1503\/cmaj.070007","volume":"176","author":"M Bance","year":"2007","unstructured":"Bance, M. Hearing and aging. CMAJ 176, 925\u2013927 (2007).","journal-title":"CMAJ"},{"key":"394_CR97","unstructured":"Community and Culture\u2014Frequently Asked Questions https:\/\/www.nad.org\/resources\/american-sign-language\/community-and-culture-frequently-asked-questions\/ (National Association of the Deaf, 2020)."},{"key":"394_CR98","doi-asserted-by":"publisher","first-page":"2381","DOI":"10.1056\/NEJMp1907970","volume":"381","author":"M Friedner","year":"2019","unstructured":"Friedner, M., Nagarajan, R., Murthy, A. & Frankfurter, R. Embracing multiple normals\u2014a 12-year-old boy in India with a cochlear implant. N. Engl. J. Med. 381, 2381\u20132384 (2019).","journal-title":"N. Engl. J. Med."}],"container-title":["Nature Machine Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00394-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00394-z","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00394-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,11]],"date-time":"2024-12-11T22:09:39Z","timestamp":1733954979000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00394-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,18]]},"references-count":98,"journal-issue":{"issue":"10","published-online":{"date-parts":[[2021,10]]}},"alternative-id":["394"],"URL":"https:\/\/doi.org\/10.1038\/s42256-021-00394-z","relation":{},"ISSN":["2522-5839"],"issn-type":[{"value":"2522-5839","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,10,18]]},"assertion":[{"value":"17 September 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 August 2021","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 October 2021","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"N.A.L. is a co-founder of Perceptual Technologies. F.-G.Z. owns stock in Axonics, Nurotron, Syntiant, Velox, Dianavi and Xense.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}