{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T19:02:25Z","timestamp":1774638145924,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":34,"publisher":"ACM","license":[{"start":{"date-parts":[[2018,10,11]],"date-time":"2018-10-11T00:00:00Z","timestamp":1539216000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2018,10,11]]},"DOI":"10.1145\/3242587.3242603","type":"proceedings-article","created":{"date-parts":[[2018,10,16]],"date-time":"2018-10-16T13:30:26Z","timestamp":1539696626000},"page":"237-246","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":51,"title":["SilentVoice"],"prefix":"10.1145","author":[{"given":"Masaaki","family":"Fukumoto","sequence":"first","affiliation":[{"name":"Microsoft Research, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2018,10,11]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"Amazon Alexa. Retrieved July 30th 2018 from https:\/\/developer.amazon.com\/alexa  Amazon Alexa. Retrieved July 30th 2018 from https:\/\/developer.amazon.com\/alexa"},{"key":"e_1_3_2_2_2_1","unstructured":"ambient::technology. Retrieved July 30th 2018 from http:\/\/wayback.archive.org\/web\/20150330030801\/http:\/\/www.theaudeo.com\/?action=technology (web archive)  ambient::technology. Retrieved July 30th 2018 from http:\/\/wayback.archive.org\/web\/20150330030801\/http:\/\/www.theaudeo.com\/?action=technology (web archive)"},{"key":"e_1_3_2_2_3_1","unstructured":"Apple Siri. Retrieved July 30th 2018 from http:\/\/www.apple.com\/iphone\/features\/siri.html  Apple Siri. Retrieved July 30th 2018 from http:\/\/www.apple.com\/iphone\/features\/siri.html"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/MC.2015.310"},{"key":"e_1_3_2_2_5_1","volume-title":"Proc. of ICASSP, 4960--4964","author":"Chan W.","unstructured":"W. Chan , N. Jaitly , Q. Le , and O. Vinyals . 2016. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition . In Proc. of ICASSP, 4960--4964 . W. Chan, N. Jaitly, Q. Le, and O. Vinyals. 2016. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition. In Proc. of ICASSP, 4960--4964."},{"key":"e_1_3_2_2_6_1","volume-title":"Addison-Wesley Professional","author":"Cohen M. H.","year":"2004","unstructured":"M. H. Cohen , J. P. Giangola , and J. Balogh . 2004. Voice user interface design . Addison-Wesley Professional , 2004 . M. H. Cohen, J. P. Giangola, and J. Balogh. 2004. Voice user interface design. Addison-Wesley Professional, 2004."},{"key":"e_1_3_2_2_7_1","unstructured":"H. Davis and S. Silverman. 1970. Hearing and Deafness. Holt Rinehart and Winston (New York).  H. Davis and S. Silverman. 1970. Hearing and Deafness. Holt Rinehart and Winston (New York)."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2009.08.002"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1017\/S0025100308003563"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1159\/000191471"},{"key":"e_1_3_2_2_11_1","unstructured":"J. S. Garofolo L. F. Lamel W. M. Fisher J. G. Fiscus D. S. Pallett and N. L. Dahlgren. 1986. The DARPA TIMIT Acoustic-Phonetic Continuous Speech Corpus CDROM. NIST Interagency\/Internal Report (NISTIR) 4930.  J. S. Garofolo L. F. Lamel W. M. Fisher J. G. Fiscus D. S. Pallett and N. L. Dahlgren. 1986. The DARPA TIMIT Acoustic-Phonetic Continuous Speech Corpus CDROM. NIST Interagency\/Internal Report (NISTIR) 4930."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2012.2205597"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2009.11.004"},{"key":"e_1_3_2_2_14_1","unstructured":"Hushme. Retrieved July 30th 2018 from http:\/\/gethushme.com\/  Hushme. Retrieved July 30th 2018 from http:\/\/gethushme.com\/"},{"key":"e_1_3_2_2_15_1","volume-title":"Development of Two Measurement Systems of Abnormal Condition by Using Condenser Microphone. Master Thesis","author":"Ishigaki T.","unstructured":"T. Ishigaki . 2005. Development of Two Measurement Systems of Abnormal Condition by Using Condenser Microphone. Master Thesis , Hosei University ( in Japanese). T. Ishigaki. 2005. Development of Two Measurement Systems of Abnormal Condition by Using Condenser Microphone. Master Thesis, Hosei University (in Japanese)."},{"key":"e_1_3_2_2_16_1","volume-title":"Proc. of ASRU'01","author":"Itoh T.","unstructured":"T. Itoh , K. Takeda , and F. Itakura . 2001. Acoustic analysis and recognition of whispered speech . In Proc. of ASRU'01 , 429--432. T. Itoh, K. Takeda, and F. Itakura. 2001. Acoustic analysis and recognition of whispered speech. In Proc. of ASRU'01, 429--432."},{"key":"e_1_3_2_2_17_1","volume-title":"Proc. of ICSLP2004","author":"Jou S. C. S.","unstructured":"S. C. S. Jou , T. Schultz , and A. Waibel . 2004. Adaptation for soft whisper recognition using a throat microphone . In Proc. of ICSLP2004 , 527--530. S. C. S. Jou, T. Schultz, and A. Waibel. 2004. Adaptation for soft whisper recognition using a throat microphone. In Proc. of ICSLP2004, 527--530."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3172944.3172977"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1049\/el.2014.1645"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/765891.765996"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1002\/scj.4690220607"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2013.10.023"},{"key":"e_1_3_2_2_23_1","volume-title":"Proc. of INTERSPEECH2015","author":"Meenakshi G. N.","unstructured":"G. N. Meenakshi , and G. P. Kumar . 2015, A discriminative analysis within and across voiced and unvoiced consonants in neutral and whispered speech in multiple indian languages . In Proc. of INTERSPEECH2015 , 781--785. G. N. Meenakshi, and G. P. Kumar. 2015, A discriminative analysis within and across voiced and unvoiced consonants in neutral and whispered speech in multiple indian languages. In Proc. of INTERSPEECH2015, 781--785."},{"key":"e_1_3_2_2_24_1","unstructured":"Microsoft Cortana. Retrieved July 30th 2018 from https:\/\/www.microsoft.com\/en-us\/windows\/cortana  Microsoft Cortana. Retrieved July 30th 2018 from https:\/\/www.microsoft.com\/en-us\/windows\/cortana"},{"key":"e_1_3_2_2_25_1","unstructured":"Microsoft Custom Speech Service. Retrieved July 30th 2018 from https:\/\/azure.microsoft.com\/en-us\/services\/cognitive-services\/custom-speech-service\/  Microsoft Custom Speech Service. Retrieved July 30th 2018 from https:\/\/azure.microsoft.com\/en-us\/services\/cognitive-services\/custom-speech-service\/"},{"key":"e_1_3_2_2_26_1","volume-title":"Proc. of ICASSP'03","author":"Nakajima Y.","unstructured":"Y. Nakajima , H. Kashioka , K. Shikano , and N. Campbell . 2003. Non-audible murmur recognition input interface using stethoscopic microphone attached to the skin . In Proc. of ICASSP'03 , 5, V-708. Y. Nakajima, H. Kashioka, K. Shikano, and N. Campbell. 2003. Non-audible murmur recognition input interface using stethoscopic microphone attached to the skin. In Proc. of ICASSP'03, 5, V-708."},{"key":"e_1_3_2_2_27_1","volume-title":"Proc. of INTERSPEECH2006","author":"Nakamura K.","unstructured":"K. Nakamura , T. Toda , H. Saruwatari , and K. Shikano . 2006. Speaking aid system for total laryngectomees using voice conversion of body transmitted artificial speech . In Proc. of INTERSPEECH2006 . http:\/\/hdl.handle.net\/10061\/8140 K. Nakamura, T. Toda, H. Saruwatari, and K. Shikano. 2006. Speaking aid system for total laryngectomees using voice conversion of body transmitted artificial speech. In Proc. of INTERSPEECH2006. http:\/\/hdl.handle.net\/10061\/8140"},{"key":"e_1_3_2_2_28_1","unstructured":"(for example) Nuance. Retrieved July 30th 2018 from https:\/\/www.nuance.com\/  (for example) Nuance. Retrieved July 30th 2018 from https:\/\/www.nuance.com\/"},{"key":"e_1_3_2_2_29_1","volume-title":"Anatomy of the cat","author":"Reighard J.","unstructured":"J. Reighard , H. S. Jennings . 1951. Anatomy of the cat . Henry Holt and Company . J. Reighard, H. S. Jennings. 1951. Anatomy of the cat. Henry Holt and Company."},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvoice.2004.10.007"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2016-1460"},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2009.11.005"},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"crossref","unstructured":"D. Wigdor and D. Wixon. 2011. Brave NUI World: Designing Natural User Interfaces for Touch and Gesture (1st ed.). Morgan Kaufmann Publishers Inc. San Francisco CA USA.   D. Wigdor and D. Wixon. 2011. Brave NUI World: Designing Natural User Interfaces for Touch and Gesture (1st ed.). Morgan Kaufmann Publishers Inc. San Francisco CA USA.","DOI":"10.1016\/B978-0-12-382231-4.00002-2"},{"key":"e_1_3_2_2_34_1","volume-title":"Proc. of ICASSP, 5255--5259","author":"Xiong W.","unstructured":"W. Xiong , J. Droppo , X. Huang , F. Seide , M. Seltzer , A. Stolcke , D. Yu , and G. Zweig . 2017. The Microsoft 2016 conversational speech recognition system . In Proc. of ICASSP, 5255--5259 . W. Xiong, J. Droppo, X. Huang, F. Seide, M. Seltzer, A. Stolcke, D. Yu, and G. Zweig. 2017. The Microsoft 2016 conversational speech recognition system. In Proc. of ICASSP, 5255--5259."}],"event":{"name":"UIST '18: The 31st Annual ACM Symposium on User Interface Software and Technology","location":"Berlin Germany","acronym":"UIST '18","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques","SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 31st Annual ACM Symposium on User Interface Software and Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3242587.3242603","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3242587.3242603","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T00:43:37Z","timestamp":1750207417000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3242587.3242603"}},"subtitle":["Unnoticeable Voice Input by Ingressive Speech"],"short-title":[],"issued":{"date-parts":[[2018,10,11]]},"references-count":34,"alternative-id":["10.1145\/3242587.3242603","10.1145\/3242587"],"URL":"https:\/\/doi.org\/10.1145\/3242587.3242603","relation":{},"subject":[],"published":{"date-parts":[[2018,10,11]]},"assertion":[{"value":"2018-10-11","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}