{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,5]],"date-time":"2025-03-05T05:39:59Z","timestamp":1741153199431,"version":"3.38.0"},"publisher-location":"Berlin, Heidelberg","reference-count":17,"publisher":"Springer Berlin Heidelberg","isbn-type":[{"type":"print","value":"9783642199165"},{"type":"electronic","value":"9783642199172"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2011]]},"DOI":"10.1007\/978-3-642-19917-2_28","type":"book-chapter","created":{"date-parts":[[2011,3,18]],"date-time":"2011-03-18T04:48:10Z","timestamp":1300423690000},"page":"233-242","source":"Crossref","is-referenced-by-count":2,"title":["Efficient Use of Voice Activity Detector and Automatic Speech Recognition in Embedded Platforms for Natural Language Interaction"],"prefix":"10.1007","author":[{"given":"Marcos","family":"Santos-P\u00e9rez","sequence":"first","affiliation":[]},{"given":"Eva","family":"Gonz\u00e1lez-Parada","sequence":"additional","affiliation":[]},{"given":"Jos\u00e9 Manuel","family":"Cano-Garc\u00eda","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"28_CR1","unstructured":"BeagleBoard website, http:\/\/beagleboard.org\/"},{"key":"28_CR2","unstructured":"CMU Sphinx website, http:\/\/cmusphinx.sourceforge.net\/"},{"key":"28_CR3","unstructured":"CMU Sphinxbase website, http:\/\/sourceforge.net\/projects\/cmusphinx\/"},{"key":"28_CR4","doi-asserted-by":"crossref","unstructured":"Huggins-daines, D., Kumar, M., Chan, A., Black, A.W., Ravishankar, M., Rudnicky, A.I.: PocketSphinx: a free, real-time continuous speech recognition system for hand-held devices. In: Proc. of ICASSP, Touluse, France, May 2006, pp. 185\u2013188 (2006)","DOI":"10.1109\/ICASSP.2006.1659988"},{"key":"28_CR5","unstructured":"Embedded ViaVoice website, http:\/\/www-01.ibm.com\/software\/pervasive\/embedded_viavoice\/"},{"issue":"4","key":"28_CR6","doi-asserted-by":"publisher","first-page":"455","DOI":"10.1016\/j.specom.2004.12.004","volume":"45","author":"H. Jiang","year":"2005","unstructured":"Jiang, H.: Confidence measures for speech recognition: A survey. Speech Communication\u00a045(4), 455\u2013470 (2005)","journal-title":"Speech Communication"},{"key":"28_CR7","doi-asserted-by":"crossref","unstructured":"Jokinen, K.: Natural language and dialogue interfaces. In: The Universal Access Handbook, 1st edn., pp. 495\u2013506. CRC Press Taylor & Francis Group (2009)","DOI":"10.1201\/9781420064995-c31"},{"key":"28_CR8","unstructured":"Loquendo ASR website, http:\/\/www.loquendo.com\/es\/technology\/asr.htm"},{"key":"28_CR9","unstructured":"Nuance VoCon website, http:\/\/www.nuance.es\/vocon\/"},{"key":"28_CR10","unstructured":"Pocketsphinx optimizations for embedded devices, http:\/\/cmusphinx.sourceforge.net\/wiki\/pocketsphinxhandhelds"},{"key":"28_CR11","doi-asserted-by":"crossref","unstructured":"Sasou, A., Kojima, H.: Noise robust speech recognition applied to Voice-Driven wheelchair. EURASIP Journal on Advances in Signal Processing 2009, 1\u201310 (2009)","DOI":"10.1155\/2009\/512314"},{"issue":"2","key":"28_CR12","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/s10772-009-9036-6","volume":"11","author":"A. Schmitt","year":"2008","unstructured":"Schmitt, A., Zaykovskiy, D., Minker, W.: Speech recognition for mobile devices. International Journal of Speech Technology\u00a011(2), 63\u201372 (2008), doi:10.1007\/s10772-009-9036-6","journal-title":"International Journal of Speech Technology"},{"key":"28_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1007\/978-3-642-12349-8_13","volume-title":"Mobile Multimedia Processing","author":"Z.-H. Tan","year":"2010","unstructured":"Tan, Z.-H., Lindberg, B.: Speech recognition on mobile devices. In: Jiang, X., Ma, M.Y., Chen, C.W. (eds.) Mobile Multimedia Processing. LNCS, vol.\u00a05960, pp. 221\u2013237. Springer, Heidelberg (2010)"},{"key":"28_CR14","unstructured":"Vertanen, K.: Baseline WSJ acoustic models for HTK and sphinx: Training recipes and recognition experiments. Technical report, University of Cambridge, Cavendish Laboratory (2006)"},{"key":"28_CR15","first-page":"237","volume-title":"IUI 2009: Proceedings of the 14th International Conference on Intelligent User Interfaces","author":"K. Vertanen","year":"2009","unstructured":"Vertanen, K., Kristensson, P.O.: Parakeet: A continuous speech recognition system for mobile Touch-Screen devices. In: IUI 2009: Proceedings of the 14th International Conference on Intelligent User Interfaces, pp. 237\u2013246. ACM Press, Sanibel Island (2009)"},{"key":"28_CR16","unstructured":"Voxforge Spanish Model website, http:\/\/cmusphinx.sourceforge.net\/2010\/08\/voxforge-spanish-model-released\/"},{"key":"28_CR17","doi-asserted-by":"crossref","unstructured":"Zhang, J., Ward, W., Pellom, B., Yu, X., Hacioglu, K.: Improvements in audio processing and language modeling in the CU communicator. In: Eurospeech 2001, Aalborg, Denmark (2001)","DOI":"10.21437\/Eurospeech.2001-521"}],"container-title":["Advances in Intelligent and Soft Computing","Highlights in Practical Applications of Agents and Multiagent Systems"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-642-19917-2_28.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,4]],"date-time":"2025-03-04T16:05:19Z","timestamp":1741104319000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-642-19917-2_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2011]]},"ISBN":["9783642199165","9783642199172"],"references-count":17,"URL":"https:\/\/doi.org\/10.1007\/978-3-642-19917-2_28","relation":{},"ISSN":["1867-5662","1867-5670"],"issn-type":[{"type":"print","value":"1867-5662"},{"type":"electronic","value":"1867-5670"}],"subject":[],"published":{"date-parts":[[2011]]}}}