{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,20]],"date-time":"2025-10-20T10:22:10Z","timestamp":1760955730992,"version":"3.37.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783319664286"},{"type":"electronic","value":"9783319664293"}],"license":[{"start":{"date-parts":[[2017,1,1]],"date-time":"2017-01-01T00:00:00Z","timestamp":1483228800000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2017]]},"DOI":"10.1007\/978-3-319-66429-3_9","type":"book-chapter","created":{"date-parts":[[2017,8,12]],"date-time":"2017-08-12T02:02:55Z","timestamp":1502503375000},"page":"98-108","source":"Crossref","is-referenced-by-count":6,"title":["An Algorithm for Detection of Breath Sounds in Spontaneous Speech with Application to Speaker Recognition"],"prefix":"10.1007","author":[{"given":"Sri Harsha","family":"Dumpala","sequence":"first","affiliation":[]},{"given":"K. N. R. K. Raju","family":"Alluri","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2017,8,13]]},"reference":[{"key":"9_CR1","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1016\/j.neucom.2014.04.002","volume":"141","author":"B Lei","year":"2014","unstructured":"Lei, B., Rahman, S.A., Song, I.: Content-based classification of breath sound with enhanced features. Neurocomputing 141, 139\u2013147 (2014)","journal-title":"Neurocomputing"},{"key":"9_CR2","doi-asserted-by":"crossref","unstructured":"Dumpala, S.H., Sridaran, K.V., Gangashetty, S.V., Yegnanarayana, B.: Analysis of laughter and speech-laugh signals using excitation source information. In: ICASSP, pp. 975\u2013979 (2014)","DOI":"10.1109\/ICASSP.2014.6853742"},{"key":"9_CR3","unstructured":"Drugman, T., Urbain, J., Dutoit, T.: Assessment of audio features for automatic cough detection. In: EUSIPCO, pp. 1289\u20131293 (2011)"},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Dumpala, S.H., Gangamohan, P., Gangashetty, S.V., Yegnanarayana, B.: Use of vowels in discriminating speech-laugh from laughter and neutral speech. In: Interspeech, pp. 1437\u20131441 (2016)","DOI":"10.21437\/Interspeech.2016-1114"},{"issue":"3","key":"9_CR5","doi-asserted-by":"crossref","first-page":"838","DOI":"10.1109\/TASL.2006.889750","volume":"15","author":"D Ruinskiy","year":"2007","unstructured":"Ruinskiy, D., Lavner, Y.: An effective algorithm for automatic detection and exact demarcation of breath sounds in speech and song signals. IEEE Trans. Audio Speech Lang. Process. 15(3), 838\u2013850 (2007)","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"9_CR6","doi-asserted-by":"crossref","unstructured":"Zelasko, P., Jadczyk, T., Zilko, B.: HMM-based breath and filled pauses elimination in ASR. In: SIGMAP, pp. 255\u2013260 (2014)","DOI":"10.5220\/0005023002550260"},{"key":"9_CR7","doi-asserted-by":"crossref","unstructured":"Igras, M., Zilko, B.: Wavelet method for breath detection in audio signals. In: ICME, pp. 1\u20136 (2013)","DOI":"10.1109\/ICME.2013.6607428"},{"key":"9_CR8","first-page":"1","volume":"1","author":"KW Godin","year":"2015","unstructured":"Godin, K.W., Hansen, J.H.: Physical task stress and speaker variability in voice quality. EURASIP J. Audio Speech Music Proc. 1, 1\u201313 (2015)","journal-title":"EURASIP J. Audio Speech Music Proc."},{"key":"9_CR9","unstructured":"Nakano, T., Ogata, J., Goto, M., Hiraga, Y.: Analysis and automatic detection of breath sounds in unaccompanied singing voice. In: ICMPC, pp. 387\u2013390 (2008)"},{"key":"9_CR10","unstructured":"Igras, M., Zilko, B.: Different types of pauses as a source of biometry. In: Models and Analysis of Vocal Emissions for Biomedical Applications, pp. 197\u2013200 (2013)"},{"key":"9_CR11","doi-asserted-by":"crossref","unstructured":"Rapcan, V., D\u2019Arcy, S., Reilly, R.B.: Automatic breath sound detection and removal for cognitive studies of speech and language. In: ISSC, pp. 1\u20136 (2009)","DOI":"10.1049\/cp.2009.1704"},{"key":"9_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1007\/978-3-642-32790-2_69","volume-title":"Text, Speech and Dialogue","author":"A Janicki","year":"2012","unstructured":"Janicki, A.: On the impact of non-speech sounds on speaker recognition. In: Sojka, P., Hor\u00e1k, A., Kope\u010dek, I., Pala, K. (eds.) TSD 2012. LNCS, vol. 7499, pp. 566\u2013572. Springer, Heidelberg (2012). doi: 10.1007\/978-3-642-32790-2_69"},{"key":"9_CR13","unstructured":"Pitt, M.A., Dilley, L., Johnson, K., Kiesling, S., Raymond, W., Hume, E., Fosler-Lussier, E.: Buckeye Corpus of Conversational Speech (2nd release). Department of Psychology, Ohio State University (Distributor), Columbus, OH (2007)"},{"key":"9_CR14","doi-asserted-by":"crossref","unstructured":"Dumpala, S.H., Nellore, B.T., Nevali, R.R., Gangashetty, S.V., Yegnanarayana, B.: Robust features for sonorant segmentation in continuous speech. In: Interspeech, pp. 1987\u20131991 (2015)","DOI":"10.21437\/Interspeech.2015-437"},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Dumpala, S.H., Nellore, B.T., Nevali, R.R., Gangashetty, S.V., Yegnanarayana, B.: Robust vowel landmark detection using epoch-based features. In: Interspeech, pp. 160\u2013164 (2016)","DOI":"10.21437\/Interspeech.2016-1074"},{"key":"9_CR16","doi-asserted-by":"crossref","first-page":"1602","DOI":"10.1109\/TASL.2008.2004526","volume":"16","author":"KSR Murty","year":"2008","unstructured":"Murty, K.S.R., Yegnanarayana, B.: Epoch extraction from speech signals. IEEE Trans. Audio Speech Lang. Process. 16, 1602\u20131613 (2008)","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"issue":"6","key":"9_CR17","doi-asserted-by":"crossref","first-page":"782","DOI":"10.1016\/j.specom.2013.02.007","volume":"55","author":"B Yegnanarayana","year":"2013","unstructured":"Yegnanarayana, B., Dhananjaya, N.G.: Spectro-temporal analysis of speech signals using zero-time windowing and group delay function. Speech Commun. 55(6), 782\u2013795 (2013)","journal-title":"Speech Commun."},{"key":"9_CR18","unstructured":"Hirose, H.: Investigating the physiology of laryngeal structures. In: The Handbook of Phonetic Sciences, Cambridge, pp. 116\u2013136 (1995)"},{"key":"9_CR19","unstructured":"Brookes, M., et al.: Voicebox: Speech processing toolbox for Matlab (2011). www.ee.ic.ac.uk\/hp\/staff\/dmb\/voicebox\/voicebox.html"},{"key":"9_CR20","unstructured":"Voice Biometry Standardization (VBS) (2015). http:\/\/voicebiometry.org\/"},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Dumpala, S.H., Kopparapu, S.K.: Improved speaker recognition system for stressed speech using deep neural networks. In: IJCNN, pp. 1257\u20131264 (2017)","DOI":"10.1109\/IJCNN.2017.7965997"}],"container-title":["Lecture Notes in Computer Science","Speech and Computer"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-319-66429-3_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T03:33:07Z","timestamp":1659324787000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-319-66429-3_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2017]]},"ISBN":["9783319664286","9783319664293"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-319-66429-3_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2017]]}}}