{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T09:04:16Z","timestamp":1743152656900,"version":"3.40.3"},"publisher-location":"Berlin, Heidelberg","reference-count":27,"publisher":"Springer Berlin Heidelberg","isbn-type":[{"type":"print","value":"9783642391873"},{"type":"electronic","value":"9783642391880"}],"license":[{"start":{"date-parts":[[2013,1,1]],"date-time":"2013-01-01T00:00:00Z","timestamp":1356998400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2013]]},"DOI":"10.1007\/978-3-642-39188-0_56","type":"book-chapter","created":{"date-parts":[[2013,7,2]],"date-time":"2013-07-02T13:03:45Z","timestamp":1372770225000},"page":"520-529","source":"Crossref","is-referenced-by-count":8,"title":["Multimodal Synthesizer for Russian and Czech Sign Languages and Audio-Visual Speech"],"prefix":"10.1007","author":[{"given":"Alexey","family":"Karpov","sequence":"first","affiliation":[]},{"given":"Zdenek","family":"Krnoul","sequence":"additional","affiliation":[]},{"given":"Milos","family":"Zelezny","sequence":"additional","affiliation":[]},{"given":"Andrey","family":"Ronzhin","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"56_CR1","unstructured":"DePaul ASL Synthesizer, http:\/\/asl.cs.depaul.edu"},{"key":"56_CR2","unstructured":"Efthimiou, E., et al.: Sign Language technologies and resources of the Dicta-Sign project. In: Proc. 5th Workshop on the Representation and Processing of Sign Languages, Istanbul, Turkey, pp. 37\u201344 (2012)"},{"key":"56_CR3","unstructured":"Dicta-Sign Project, www.dictasign.eu"},{"key":"56_CR4","unstructured":"Caminero, J., Rodr\u00edguez-Gancedo, M., Hern\u00e1ndez-Trapote, A., L\u00f3pez-Menc\u00eda, B.: SIGNSPEAK Project Tools: A way to improve the communication bridge between signer and hearing communities. In: Proc. 5th Workshop on the Representation and Processing of Sign Languages, Istanbul, Turkey, pp. 1\u20136 (2012)"},{"key":"56_CR5","unstructured":"SIGNSPEAK Project, www.signspeak.eu\/en"},{"key":"56_CR6","doi-asserted-by":"crossref","unstructured":"Gibet, S., Courty, N., Duarte, K., Naour, T.: The SignCom system for data-driven animation of interactive virtual signers: Methodology and Evaluation. ACM Transactions on Interactive Intelligent Systems\u00a01(1) (2011)","DOI":"10.1145\/2030365.2030371"},{"key":"56_CR7","unstructured":"Borgotallo, R., et al.: A multi-language database for supporting sign language translation and synthesis. In: Proc. 4th Workshop on the Representation and Processing of Sign Languages: Corpora and Sign Language Technologies, Malta, pp. 23\u201326 (2010)"},{"key":"56_CR8","unstructured":"ViSiCAST Project, www.visicast.co.uk"},{"key":"56_CR9","unstructured":"eSign Project, www.sign-lang.uni-hamburg.de\/esign"},{"key":"56_CR10","unstructured":"Vcom3D Company, www.vcom3d.com"},{"key":"56_CR11","unstructured":"SiSi Project, www-03.ibm.com\/press\/us\/en\/pressrelease\/22316.wss"},{"key":"56_CR12","unstructured":"iCommunicator project, www.icommunicator.com"},{"issue":"7","key":"56_CR13","doi-asserted-by":"publisher","first-page":"335","DOI":"10.1023\/B:IJST.0000037076.86366.8d","volume":"4","author":"J. Beskow","year":"2004","unstructured":"Beskow, J.: Trainable articulatory control models for visual speech synthesis. Journal of Speech Technology\u00a04(7), 335\u2013349 (2004)","journal-title":"Journal of Speech Technology"},{"key":"56_CR14","doi-asserted-by":"crossref","unstructured":"Youssef, A., Hueber, T., Badin, P., Bailly, G.: Toward a Multi-Speaker Visual Articulatory Feedback System. In: Proc. International Conference INTERSPEECH-2011, Florence, Italy, pp. 589\u2013592 (2011)","DOI":"10.21437\/Interspeech.2011-238"},{"key":"56_CR15","unstructured":"Hanke, T.: HamNoSys - Representing sign language data in language resources and language processing contexts. In: Proc. International Conference on Language Resources and Evaluation LREC-2004, Lisbon, Portugal, pp. 1\u20136 (2004)"},{"key":"56_CR16","unstructured":"Hoffmann, R., Jokisch, O., Lobanov, B., Tsirulnik, L., Shpilewsky, E., Piurkowska, B., Ronzhin, A., Karpov, A.: Slavonic TTS and SST Conversion for Let\u2019s Fly Dialogue System. In: Proc. 12th International Conference on Speech and Computer SPECOM-2007, Moscow, Russia, pp. 729\u2013733 (2007)"},{"key":"56_CR17","doi-asserted-by":"crossref","unstructured":"Tihelka, D., Kala, J., Matou\u0161ek, J.: Enhancements of Viterbi Search for Fast Unit Selection Synthesis. In: Proc. International Conference INTERSPEECH-2010, Makuhari, Japan, pp. 174\u2013177 (2010)","DOI":"10.21437\/Interspeech.2010-78"},{"issue":"12","key":"56_CR18","doi-asserted-by":"publisher","first-page":"3657","DOI":"10.1016\/j.sigpro.2006.02.039","volume":"86","author":"M. \u017delezn\u00fd","year":"2006","unstructured":"\u017delezn\u00fd, M., Kr\u0148oul, Z., Cisar, P., Matousek, J.: Design, implementation and evaluation of the Czech realistic audio-visual speech synthesis. Signal Processing\u00a086(12), 3657\u20133673 (2006)","journal-title":"Signal Processing"},{"key":"56_CR19","doi-asserted-by":"crossref","unstructured":"Karpov, A., Tsirulnik, L., Kr\u0148oul, Z., Ronzhin, A., Lobanov, B., \u017delezn\u00fd, M.: Audio-Visual Speech Asynchrony Modeling in a Talking Head. In: Proc. International Conference INTERSPEECH-2009, Brighton, UK, pp. 2911\u20132914 (2009)","DOI":"10.21437\/Interspeech.2009-737"},{"key":"56_CR20","series-title":"Lecture Notes in Computer Science","first-page":"180","volume-title":"Machine Learning for Multimodal Interaction","author":"Z. Kr\u0148oul","year":"2007","unstructured":"Kr\u0148oul, Z., Kanis, J., \u017delezn\u00fd, M., M\u00fcller, L.: Czech Text-to-Sign Speech Synthesizer. In: Popescu-Belis, A., Renals, S., Bourlard, H. (eds.) MLMI 2007. LNCS, vol.\u00a04892, pp. 180\u2013191. Springer, Heidelberg (2007)"},{"key":"56_CR21","doi-asserted-by":"crossref","unstructured":"Kr\u0148oul, Z., \u017delezn\u00fd, M., M\u00fcller, L.: Training of Coarticulation Models using Dominance Functions and Visual Unit Selection Methods for Audio-Visual Speech Synthesis. In: Proc. 9th International Conference on Spoken Language Processing INTERSPEECH-2006, Pittsburgh, PA, pp. 585\u2013588 (2006)","DOI":"10.21437\/Interspeech.2006-215"},{"key":"56_CR22","unstructured":"Karpov, A., Ronzhin, A., Kipyatkova, I., \u017delezn\u00fd, M.: Influence of Phone-viseme Temporal Correlations on Audiovisual STT and TTS Performance. In: Proc. 17th International Congress of Phonetic Sciences ICPhS-2011, Hong Kong, China, pp. 1030\u20131033 (2011)"},{"key":"56_CR23","series-title":"LNAI","doi-asserted-by":"publisher","first-page":"613","DOI":"10.1007\/11846406_77","volume-title":"Text, Speech and Dialogue","author":"J. Kanis","year":"2006","unstructured":"Kanis, J., Zahradil, J., Jur\u010d\u00ed\u010dek, F., M\u00fcller, L.: Czech-Sign Speech corpus for semantic based machine translation. In: Sojka, P., Kope\u010dek, I., Pala, K. (eds.) TSD 2006. LNCS (LNAI), vol.\u00a04188, pp. 613\u2013620. Springer, Heidelberg (2006)"},{"issue":"2","key":"56_CR24","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1007\/s12193-011-0059-3","volume":"4","author":"M. Hr\u00faz","year":"2011","unstructured":"Hr\u00faz, M., Campr, P., Dikici, E., Kindirouglu, A., Kr\u0148oul, Z., Ronzhin, A., Sak, H., Schorno, D., Akarun, L., Aran, O., Karpov, A., Saraclar, M., \u017delezn\u00fd, M.: Automatic Fingersign to Speech Translation System. Journal on Multimodal User Interfaces\u00a04(2), 61\u201379 (2011)","journal-title":"Journal on Multimodal User Interfaces"},{"key":"56_CR25","doi-asserted-by":"crossref","unstructured":"Kr\u0148oul, Z.: Web-based sign language synthesis and animation for on-line assistive technologies. In: Proc. 13th International ACM SIGACCESS Conference on Computers and Accessibility ASSETS-2011, Dundee, Scotland, UK, pp. 307\u2013308 (2011)","DOI":"10.1145\/2049536.2049620"},{"key":"56_CR26","unstructured":"Audio-visual demonstration of the universal multimodal synthesizer for Russian, http:\/\/www.spiiras.nw.ru\/speech\/demo\/daktilrus.avi"},{"key":"56_CR27","series-title":"LNAI","doi-asserted-by":"publisher","first-page":"524","DOI":"10.1007\/978-3-540-74628-7_68","volume-title":"Text, Speech and Dialogue","author":"Z. Kr\u0148oul","year":"2007","unstructured":"Kr\u0148oul, Z., \u017delezn\u00fd, M.: Translation and conversion for Czech Sign Speech synthesis. In: Matou\u0161ek, V., Mautner, P. (eds.) TSD 2007. LNCS (LNAI), vol.\u00a04629, pp. 524\u2013531. Springer, Heidelberg (2007)"}],"container-title":["Lecture Notes in Computer Science","Universal Access in Human-Computer Interaction. Design Methods, Tools, and Interaction Techniques for eInclusion"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-642-39188-0_56","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,2,27]],"date-time":"2022-02-27T01:15:04Z","timestamp":1645924504000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-642-39188-0_56"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2013]]},"ISBN":["9783642391873","9783642391880"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-642-39188-0_56","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2013]]}}}