{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T06:12:05Z","timestamp":1725862325615},"publisher-location":"Cham","reference-count":24,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783319401737"},{"type":"electronic","value":"9783319401744"}],"license":[{"start":{"date-parts":[[2016,8,6]],"date-time":"2016-08-06T00:00:00Z","timestamp":1470441600000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2017]]},"DOI":"10.1007\/978-3-319-40174-4_6","type":"book-chapter","created":{"date-parts":[[2016,8,5]],"date-time":"2016-08-05T06:21:15Z","timestamp":1470378075000},"page":"93-99","source":"Crossref","is-referenced-by-count":0,"title":["Conclusions"],"prefix":"10.1007","author":[{"given":"Jo\u00e3o","family":"Freitas","sequence":"first","affiliation":[]},{"given":"Ant\u00f3nio","family":"Teixeira","sequence":"additional","affiliation":[]},{"given":"Miguel Sales","family":"Dias","sequence":"additional","affiliation":[]},{"given":"Samuel","family":"Silva","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2016,8,6]]},"reference":[{"key":"6_CR1","unstructured":"Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M (2016) TensorFlow: Large-scale machine learning on heterogeneous distributed systems, arXiv preprint arXiv:1603.04467"},{"key":"6_CR2","doi-asserted-by":"crossref","unstructured":"Alghowinem, S, Wagner, M, Goecke, R (2013) AusTalk\u2014The Australian speech database: design framework, recording experience and localisation. In: 8th Int. Conf. on Information Technology in Asia (CITA 2013). IEEE, pp 1\u20137","DOI":"10.1109\/CITA.2013.6637567"},{"key":"6_CR3","doi-asserted-by":"crossref","first-page":"173","DOI":"10.1016\/j.csl.2015.05.005","volume":"36","author":"L Badino","year":"2016","unstructured":"Badino L, Canevari C, Fadiga L, Metta G (2016) Integrating articulatory data in deep neural network-based acoustic modeling. Comput Speech Lang 36:173\u2013195","journal-title":"Comput Speech Lang"},{"key":"6_CR4","doi-asserted-by":"crossref","unstructured":"Barga R, Fontama V, Tok WH (2015) Introducing Microsoft Azure Machine Learning. In: Predictive analytics with Microsoft Azure Machine Learning. Springer, New York, pp 21\u201343","DOI":"10.1007\/978-1-4842-1200-4_2"},{"key":"6_CR5","doi-asserted-by":"crossref","unstructured":"Bedri A, Byrd D, Presti P, Sahni H, Gue Z, Starner T (2015a) Stick it in your ear: building an in-ear jaw movement sensor. In: Proceedings of the 2015 ACM international joint conference on pervasive and ubiquitous computing and proceedings of the 2015 ACM international symposium on wearable computers, ACM, pp 1333\u20131338","DOI":"10.1145\/2800835.2807933"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Bedri A, Sahni H, Thukral P, Starner T, Byrd D, Presti P, Reyes G, Ghovanloo M, Guo Z, (2015b) Toward silent-speech control of consumer wearables. Computer (Long Beach Calif) 54\u201362","DOI":"10.1109\/MC.2015.310"},{"key":"6_CR7","doi-asserted-by":"crossref","unstructured":"Bocquelet F, Hueber T, Girin L, Savariaux C, Yvert B (2015) Real-time control of a DNN-based articulatory synthesizer for silent speech conversion: a pilot study. In: Sixteenth annual conference of the international speech communication association","DOI":"10.21437\/Interspeech.2015-520"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Cheah LA, Gilbert JM, Gonzalez JA, Bai J, Ell SR, Fagan MJ, Moore RK, Green PD, Rychenko SI (2015) Integrating user-centred design in the development of a silent speech interface based on permanent magnetic articulography. In: Biomedical engineering systems and technologies. Springer, Berlin, pp 324\u2013337","DOI":"10.1007\/978-3-319-27707-3_20"},{"key":"6_CR9","doi-asserted-by":"crossref","unstructured":"Copeland M, Soh J, Puca A, Manning M, Gollob D (2015) Microsoft Azure Machine Learning. In: Microsoft Azure. Springer, New York, pp 355\u2013380","DOI":"10.1007\/978-1-4842-1043-7_14"},{"key":"6_CR10","doi-asserted-by":"publisher","DOI":"10.1007\/s12193-013-0120-5","author":"DA Dahl","year":"2013","unstructured":"Dahl DA (2013) The W3C multimodal architecture and interfaces standard. J Multimodal User Interfaces. doi: 10.1007\/s12193-013-0120-5","journal-title":"J Multimodal User Interfaces"},{"key":"6_CR11","doi-asserted-by":"publisher","unstructured":"Diener L, Janke M, Schultz T (2015) Direct conversion from facial myoelectric signals to speech using Deep Neural Networks. In: Neural Networks (IJCNN), 2015 Int. Jt. Conf. doi: 10.1109\/IJCNN.2015.7280404","DOI":"10.1109\/IJCNN.2015.7280404"},{"key":"6_CR12","unstructured":"Freitas J, Candeias S, Dias MS, Lleida E, Ortega A, Teixeira A, Silva S, Acarturk C, Orvalho V (2014a) The IRIS Project: a liaison between industry and academia towards natural multimodal communication. In: Iberspeech 2014"},{"key":"6_CR13","unstructured":"Freitas J, Teixeira A, Dias MS (2014b) Multimodal Corpora for Silent Speech Interaction. In: 9th Language resources and evaluation conference, pp 1\u20135"},{"key":"6_CR14","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0127040","volume":"10","author":"J Freitas","year":"2015","unstructured":"Freitas J, Teixeira A, Silva S, Oliveira C, Dias MS (2015) Detecting Nasal Vowels in Speech Interfaces Based on Surface Electromyography. PLoS One 10, e0127040. doi: 10.1371\/journal.pone.0127040","journal-title":"PLoS One"},{"key":"6_CR15","doi-asserted-by":"publisher","unstructured":"Galatas G, Potamianos G, Makedon F (2012) Audio-visual speech recognition using depth information from the Kinect in noisy video condition. In: Proceedings of the 5th International conference on PErvasive Technologies Related to Assistive Environments\u2014PETRA\u201912, pp 1\u20134. doi: 10.1145\/2413097.2413100","DOI":"10.1145\/2413097.2413100"},{"key":"6_CR16","unstructured":"Hahm S, Wang J (2015) Silent speech recognition from articulatory movements using deep neural network. In: Proc. of the International congress of phonetic sciences"},{"key":"6_CR17","doi-asserted-by":"crossref","first-page":"274","DOI":"10.1016\/j.csl.2015.03.005","volume":"36","author":"T Hueber","year":"2016","unstructured":"Hueber T, Bailly G (2016) Statistical conversion of silent articulation into audible speech using full-covariance HMM. Comput Speech Lang 36:274\u2013293","journal-title":"Comput Speech Lang"},{"key":"6_CR18","doi-asserted-by":"crossref","unstructured":"Li W (2016) Silent speech interface design methodology and case study. Chinese J Electron 25","DOI":"10.1049\/cje.2016.01.014"},{"key":"6_CR20","doi-asserted-by":"publisher","unstructured":"Silva S, Almeida N, Pereira C, Martins AI, Rosa AF, e Silva MO, Teixeira A (2015) Design and development of multimodal applications: a vision on key issues and methods, Lecture notes in computer science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). doi: 10.1007\/978-3-319-20678-3_11","DOI":"10.1007\/978-3-319-20678-3_11"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Telaar D, Wand M, Gehrig D, Putze F, Amma C, Heger D, Vu NT, Erhardt M, Schlippe T, Janke M (2014) BioKIT-Real-time decoder for biosignal processing. In: The 15th Annual conference of the international speech communication association (Interspeech 2014)","DOI":"10.21437\/Interspeech.2014-567"},{"key":"6_CR22","unstructured":"Wand M, Schulte C, Janke M, Schultz T (2013) Array-based Electromyographic Silent Speech Interface. In: International Conference on bio-inspired systems and signal processing (BIOSIGNALS 2013)"},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Wand M, Koutn\u00edk J, Schmidhuber J (2016) Lipreading with long short-term memory. arXiv Prepr. arXiv1601.08188","DOI":"10.1109\/ICASSP.2016.7472852"},{"key":"6_CR24","doi-asserted-by":"crossref","unstructured":"Wang J, Hahm S, Mau T (2015) Determining an optimal set of flesh points on tongue, lips, and jaw for continuous silent speech recognition. In: 6th Workshop on speech and language processing for assistive technologies (SLPAT), p 79","DOI":"10.18653\/v1\/W15-5114"},{"key":"6_CR25","first-page":"285","volume":"8","author":"H Yamaguchi","year":"2015","unstructured":"Yamaguchi H, Yamazaki T, Yamamoto K, Ueno S, Yamaguchi A, Ito T, Hirose S, Kamijo K, Takayanagi H, Yamanoi T (2015) Decoding silent speech in Japanese from single trial EEGS: preliminary results. J Comput Sci Syst Biol 8:285","journal-title":"J Comput Sci Syst Biol"}],"container-title":["SpringerBriefs in Electrical and Computer Engineering","An Introduction to Silent Speech Interfaces"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-319-40174-4_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,5]],"date-time":"2022-07-05T05:06:27Z","timestamp":1656997587000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-319-40174-4_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2016,8,6]]},"ISBN":["9783319401737","9783319401744"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-319-40174-4_6","relation":{},"ISSN":["2191-8112","2191-8120"],"issn-type":[{"type":"print","value":"2191-8112"},{"type":"electronic","value":"2191-8120"}],"subject":[],"published":{"date-parts":[[2016,8,6]]}}}