{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T06:35:34Z","timestamp":1763620534058,"version":"3.40.3"},"publisher-location":"Cham","reference-count":65,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030122089"},{"type":"electronic","value":"9783030122096"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-12209-6_15","type":"book-chapter","created":{"date-parts":[[2019,1,23]],"date-time":"2019-01-23T23:09:43Z","timestamp":1548284983000},"page":"305-328","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Optimizing Phoneme-to-Viseme Mapping for Continuous Lip-Reading in Spanish"],"prefix":"10.1007","author":[{"given":"Adriana","family":"Fernandez-Lopez","sequence":"first","affiliation":[]},{"given":"Federico M.","family":"Sukno","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,1,23]]},"reference":[{"key":"15_CR1","doi-asserted-by":"publisher","first-page":"746","DOI":"10.1038\/264746a0","volume":"264","author":"H McGurk","year":"1976","unstructured":"McGurk, H., MacDonald, J.: Hearing lips and seeing voices. Nature 264, 746\u2013748 (1976)","journal-title":"Nature"},{"key":"15_CR2","doi-asserted-by":"publisher","first-page":"1306","DOI":"10.1109\/JPROC.2003.817150","volume":"91","author":"G Potamianos","year":"2003","unstructured":"Potamianos, G., Neti, C., Gravier, G., Garg, A., Senior, A.W.: Recent advances in the automatic recognition of audiovisual speech. Proc. IEEE 91, 1306\u20131326 (2003)","journal-title":"Proc. IEEE"},{"key":"15_CR3","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1111\/j.1467-8721.2008.00615.x","volume":"17","author":"LD Rosenblum","year":"2008","unstructured":"Rosenblum, L.D.: Speech perception as a multimodal phenomenon. Curr. Dir. Psychol. Sci. 17, 405\u2013409 (2008)","journal-title":"Curr. Dir. Psychol. Sci."},{"key":"15_CR4","doi-asserted-by":"crossref","unstructured":"Chi\u0163u, A., Rothkrantz, L.J.: Automatic visual speech recognition. In: Speech Enhancement, Modeling and Recognition-Algorithms and Applications, pp. 95\u2013120 (2012)","DOI":"10.5772\/36466"},{"key":"15_CR5","doi-asserted-by":"publisher","first-page":"481","DOI":"10.1044\/jshd.4004.481","volume":"40","author":"NP Erber","year":"1975","unstructured":"Erber, N.P.: Auditory-visual perception of speech. J. Speech Hear. Disord. 40, 481\u2013492 (1975)","journal-title":"J. Speech Hear. Disord."},{"key":"15_CR6","doi-asserted-by":"publisher","first-page":"212","DOI":"10.1121\/1.1907309","volume":"26","author":"WH Sumby","year":"1954","unstructured":"Sumby, W.H., Pollack, I.: Visual contribution to speech intelligibility in noise. J. Acoust. Soc. Am. 26, 212\u2013215 (1954)","journal-title":"J. Acoust. Soc. Am."},{"key":"15_CR7","unstructured":"Hilder, S., Harvey, R., Theobald, B.J.: Comparison of human and machine-based lip-reading. In: Proceedings of Conference on Auditory-Visual Speech Processing, pp. 86\u201389 (2009)"},{"key":"15_CR8","doi-asserted-by":"publisher","first-page":"1601","DOI":"10.3758\/APP.72.6.1601","volume":"72","author":"RE Ronquest","year":"2010","unstructured":"Ronquest, R.E., Levi, S.V., Pisoni, D.B.: Language identification from visual-only speech signals. Atten. Percept. Psychophys. 72, 1601\u20131613 (2010)","journal-title":"Atten. Percept. Psychophys."},{"key":"15_CR9","first-page":"14","volume":"2008","author":"R Seymour","year":"2008","unstructured":"Seymour, R., Stewart, D., Ming, J.: Comparison of image transform-based features for visual speech recognition in clean and corrupted videos. J. Image Video Process. 2008, 14 (2008)","journal-title":"J. Image Video Process."},{"key":"15_CR10","doi-asserted-by":"crossref","unstructured":"Antonakos, E., Roussos, A., Zafeiriou, S.: A survey on mouth modeling and analysis for sign language recognition. In: Proceedings of Conference on Automatic Face and Gesture Recognition, vol. 1, pp. 1\u20137 (2015)","DOI":"10.1109\/FG.2015.7163162"},{"key":"15_CR11","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1109\/6046.865479","volume":"2","author":"S Dupont","year":"2000","unstructured":"Dupont, S., Luettin, J.: Audio-visual speech modeling for continuous speech recognition. IEEE Trans. Multimedia 2, 141\u2013151 (2000)","journal-title":"IEEE Trans. Multimedia"},{"key":"15_CR12","doi-asserted-by":"crossref","unstructured":"Nefian, A.V., Liang, L., Pi, X., Xiaoxiang, L., Mao, C., Murphy, K.: A coupled HMM for audio-visual speech recognition. In: Proceedings of Conference on Acoustics, Speech, and Signal Processing, vol. 2, pp. 2013\u20132016 (2002)","DOI":"10.1109\/ICASSP.2002.1006167"},{"key":"15_CR13","doi-asserted-by":"publisher","first-page":"590","DOI":"10.1016\/j.imavis.2014.06.004","volume":"32","author":"Z Zhou","year":"2014","unstructured":"Zhou, Z., Zhao, G., Hong, X., Pietik\u00e4inen, M.: A review of recent advances in visual speech decoding. Image Vis. Comput. 32, 590\u2013605 (2014)","journal-title":"Image Vis. Comput."},{"key":"15_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"832","DOI":"10.1007\/978-3-540-74272-2_103","volume-title":"Computer Analysis of Images and Patterns","author":"WC Yau","year":"2007","unstructured":"Yau, W.C., Kumar, D.K., Weghorn, H.: Visual speech recognition using motion features and hidden Markov models. In: Kropatsch, W.G., Kampel, M., Hanbury, A. (eds.) CAIP 2007. LNCS, vol. 4673, pp. 832\u2013839. Springer, Heidelberg (2007). https:\/\/doi.org\/10.1007\/978-3-540-74272-2_103"},{"key":"15_CR15","doi-asserted-by":"crossref","unstructured":"Sui, C., Bennamoun, M., Togneri, R.: Listening with your eyes: towards a practical visual speech recognition system using deep Boltzmann machines. In: Proceedings of IEEE International Conference on Computer Vision, pp. 154\u2013162 (2015)","DOI":"10.1109\/ICCV.2015.26"},{"key":"15_CR16","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Senior, A., Vinyals, O., Zisserman, A.: Lip reading sentences in the wild. In: Proceedings of Conference on Computer Vision and Pattern Recognition, pp. 3444\u20133453 (2017)","DOI":"10.1109\/CVPR.2017.367"},{"key":"15_CR17","doi-asserted-by":"crossref","unstructured":"Petridis, S., Pantic, M.: Deep complementary bottleneck features for visual speech recognition. In: Proceedings of Conference on Acoustics, Speech and Signal Processing, pp. 2304\u20132308 (2016)","DOI":"10.1109\/ICASSP.2016.7472088"},{"key":"15_CR18","doi-asserted-by":"crossref","unstructured":"Almajai, I., Cox, S., Harvey, R., Lan, Y.: Improved speaker independent lip reading using speaker adaptive training and deep neural networks. In: Proceedings of Conference on Acoustics, Speech and Signal Processing, pp. 2722\u20132726 (2016)","DOI":"10.1109\/ICASSP.2016.7472172"},{"key":"15_CR19","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1007\/978-3-319-54184-6_6","volume-title":"Computer Vision \u2013 ACCV 2016","author":"Joon Son Chung","year":"2017","unstructured":"Chung, J.S., Zisserman, A.: Lip reading in the wild. In: Proceedings of Asian Conference on Computer Vision, pp. 87\u2013103 (2016)"},{"key":"15_CR20","doi-asserted-by":"crossref","unstructured":"Wand, M., Koutn\u00edk, J., Schmidhuber, J.: Lipreading with long short-term memory. In: Proceedings Conference on Acoustics, Speech and Signal Processing, pp. 6115\u20136119 (2016)","DOI":"10.1109\/ICASSP.2016.7472852"},{"key":"15_CR21","doi-asserted-by":"publisher","first-page":"5","DOI":"10.2307\/522070","volume":"11","author":"WF Twaddell","year":"1935","unstructured":"Twaddell, W.F.: On defining the phoneme. Language 11, 5\u201362 (1935)","journal-title":"Language"},{"key":"15_CR22","doi-asserted-by":"publisher","first-page":"796","DOI":"10.1044\/jshr.1104.796","volume":"11","author":"CG Fisher","year":"1968","unstructured":"Fisher, C.G.: Confusions among visually perceived consonants. J. Speech Lang. Hear. Res. 11, 796\u2013804 (1968)","journal-title":"J. Speech Lang. Hear. Res."},{"key":"15_CR23","doi-asserted-by":"publisher","first-page":"678","DOI":"10.1121\/1.1912683","volume":"50","author":"KL Moll","year":"1971","unstructured":"Moll, K.L., Daniloff, R.G.: Investigation of the timing of velar movements during speech. J. Acoust. Soc. Am. 50, 678\u2013684 (1971)","journal-title":"J. Acoust. Soc. Am."},{"key":"15_CR24","unstructured":"Luettin, J., Thacker, N.A., Beet, S.W.: Visual speech recognition using active shape models and hidden Markov models. In: Proceedings of Conference on Acoustics, Speech, and Signal Processing, vol. 2, pp. 817\u2013820 (1996)"},{"key":"15_CR25","doi-asserted-by":"crossref","unstructured":"Sahu, V., Sharma, M.: Result based analysis of various lip tracking systems. In: Proceedings of Conference on Green High Performance Computing, pp. 1\u20137 (2013)","DOI":"10.1109\/ICGHPC.2013.6533911"},{"key":"15_CR26","unstructured":"Cappelletta, L., Harte, N.: Viseme definitions comparison for visual-only speech recognition. In: Proceedings of Conference on Signal Processing, pp. 2109\u20132113 (2011)"},{"key":"15_CR27","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Zisserman, A.: Lip reading in profile. In: Proceedings of British Machine Vision Conference (2017)","DOI":"10.1007\/978-3-319-54184-6_6"},{"key":"15_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"230","DOI":"10.1007\/978-3-319-14364-4_22","volume-title":"Advances in Visual Computing","author":"HL Bear","year":"2014","unstructured":"Bear, H.L., Harvey, R.W., Theobald, B.-J., Lan, Y.: Which phoneme-to-viseme maps best improve visual-only computer lip-reading? In: Bebis, G., et al. (eds.) ISVC 2014. LNCS, vol. 8888, pp. 230\u2013239. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-14364-4_22"},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Hazen, T.J., Saenko, K., La, C.H., Glass, J.R.: A segment-based audio-visual speech recognizer: data collection, development, and initial experiments. In: Proceedings of Conference on Multimodal Interfaces, pp. 235\u2013242 (2004)","DOI":"10.1145\/1027933.1027972"},{"key":"15_CR30","unstructured":"Neti, C., et al.: Audio visual speech recognition. Technical report, IDIAP (2000)"},{"key":"15_CR31","unstructured":"Goldschen, A.J., Garcia, O.N., Petajan, E.: Continuous optical automatic speech recognition by lipreading. In: Proceedings of Conference on Signals, Systems and Computers, vol. 1, pp. 572\u2013577 (1994)"},{"key":"15_CR32","volume-title":"Speechreading (Lipreading)","author":"J Jeffers","year":"1980","unstructured":"Jeffers, J., Barley, M.: Speechreading (Lipreading). Charles C. Thomas Publisher, Springfield (1980)"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Bozkurt, E., Erdem, C.E., Erzin, E., Erdem, T., Ozkan, M.: Comparison of phoneme and viseme based acoustic units for speech driven realistic lip animation. In: Proceedings of Signal Processing and Communications Applications, pp. 1\u20134 (2007)","DOI":"10.1109\/SIU.2007.4298572"},{"key":"15_CR34","unstructured":"Ezzat, T., Poggio, T.: MikeTalk: a talking facial display based on morphing visemes. In: Proceedings of Conference on Computer Animation, pp. 96\u2013102 (1998)"},{"key":"15_CR35","doi-asserted-by":"crossref","unstructured":"Fernandez-Lopez, A., Sukno, F.M.: Automatic viseme vocabulary construction to enhance continuous lip-reading. In: Proceedings of Conference on Computer Vision Theory and Applications, vol. 5, pp. 52\u201363 (2017)","DOI":"10.5220\/0006102100520063"},{"key":"15_CR36","unstructured":"Messer, K., Matas, J., Kittler, J., Luettin, J., Maitre, G.: XM2VTSDB: the extended M2VTS database. In: Proceedings of Conference on Audio and Video-Based Biometric Person Authentication, vol. 964, pp. 965\u2013966 (1999)"},{"key":"15_CR37","doi-asserted-by":"publisher","first-page":"1145","DOI":"10.1109\/TASL.2011.2172427","volume":"20","author":"V Estellers","year":"2012","unstructured":"Estellers, V., Gurban, M., Thiran, J.P.: On dynamic stream weighting for audio-visual speech recognition. IEEE-ACM Trans. Audio Speech Lang. Process. 20, 1145\u20131157 (2012)","journal-title":"IEEE-ACM Trans. Audio Speech Lang. Process."},{"key":"15_CR38","first-page":"1","volume-title":"Multimodal Processing and Interaction","author":"George Papandreou","year":"2008","unstructured":"Papandreou, G., Katsamanis, A., Pitsikalis, V., Maragos, P.: Adaptive multimodal fusion by uncertainty compensation with application to audio-visual speech recognition. In: Proceedings of Multimodal Processing and Interaction, pp. 1\u201315 (2008)"},{"key":"15_CR39","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1109\/TASL.2008.2011515","volume":"17","author":"G Papandreou","year":"2009","unstructured":"Papandreou, G., Katsamanis, A., Pitsikalis, V., Maragos, P.: Adaptive multimodal fusion by uncertainty compensation with application to audiovisual speech recognition. IEEE-ACM Trans. Audio Speech Lang. Process. 17, 423\u2013435 (2009)","journal-title":"IEEE-ACM Trans. Audio Speech Lang. Process."},{"key":"15_CR40","unstructured":"Estival, D., Cassidy, S., Cox, F., Burnham, D.: AusTalk: an audio-visual corpus of Australian English. In: Proceeding of Conference on Language Resources and Evaluation (2014)"},{"key":"15_CR41","unstructured":"Ukai, N., Seko, T., Tamura, S., Hayamizu, S.: GIF-LR: GA-based informative feature for lipreading. In: Proceedings Conference on Signal and Information Processing Association Annual Summit and Conference, pp. 1\u20134 (2012)"},{"key":"15_CR42","doi-asserted-by":"crossref","unstructured":"Ninomiya, H., Kitaoka, N., Tamura, S., Iribe, Y., Takeda, K.: Integration of deep bottleneck features for audio-visual speech recognition. In: Proceedings of Interspeech, pp. 563\u2013567 (2015)","DOI":"10.21437\/Interspeech.2015-204"},{"key":"15_CR43","unstructured":"Tamura, S., et al.: CENSREC-1-AV: an audio-visual corpus for noisy bimodal speech recognition. In: Proceedings of Conference on Auditory-Visual Speech Processing (2010)"},{"key":"15_CR44","unstructured":"Lan, Y., Harvey, R., Theobald, B., Ong, E.J., Bowden, R.: Comparing visual features for lipreading. In: Proceedings of Conference on Auditory-Visual Speech Processing, pp. 102\u2013106 (2009)"},{"key":"15_CR45","unstructured":"Kolossa, D., Zeiler, S., Vorwerk, A., Orglmeister, R.: Audiovisual speech recognition with missing or unreliable data. In: Proceedings of Conference on Auditory-Visual Speech Processing, pp. 117\u2013122 (2009)"},{"key":"15_CR46","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/TMM.2009.2030637","volume":"11","author":"G Zhao","year":"2009","unstructured":"Zhao, G., Barnard, M., Pietikainen, M.: Lipreading with local spatiotemporal descriptors. IEEE Trans. Multimed. 11, 1254\u20131265 (2009)","journal-title":"IEEE Trans. Multimed."},{"key":"15_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TPAMI.2013.173","volume":"36","author":"Z Zhou","year":"2014","unstructured":"Zhou, Z., Hong, X., Zhao, G., Pietik\u00e4inen, M.: A compact representation of visual speech data using latent variables. IEEE Trans. Pattern Anal. Mach. Intell. 36, 1 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR48","doi-asserted-by":"crossref","unstructured":"Pei, Y., Kim, T.K., Zha, H.: Unsupervised random forest manifold alignment for lipreading. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 129\u2013136 (2013)","DOI":"10.1109\/ICCV.2013.23"},{"key":"15_CR49","unstructured":"Assael, Y.M., Shillingford, B., Whiteson, S., de Freitas, N.: LipNet: sentence-level lipreading. In: Proceedings of GPU Technology Conference (2017)"},{"key":"15_CR50","unstructured":"Thangthai, K., Harvey, R., Cox, S., Theobald, B.J.: Improving lip-reading performance for robust audiovisual speech recognition using DNNs. In: Proceedings of Conference on Auditory-Visual Speech Processing (2015)"},{"key":"15_CR51","doi-asserted-by":"crossref","unstructured":"Noda, K., Yamaguchi, Y., Nakadai, K., Okuno, H.G., Ogata, T.: Lipreading using convolutional neural network. In: Proceedings of Interspeech, pp. 1149\u20131153 (2014)","DOI":"10.21437\/Interspeech.2014-293"},{"key":"15_CR52","doi-asserted-by":"crossref","unstructured":"Lan, Y., Harvey, R., Theobald, B.J.: Insights into machine lip reading. In: Proceedings of Conference on Acoustics, Speech and Signal Processing, pp. 4825\u20134828 (2012)","DOI":"10.1109\/ICASSP.2012.6288999"},{"key":"15_CR53","unstructured":"Ortega, A., et al.: AV@CAR: a Spanish multichannel multimodal corpus for in-vehicle automatic audio-visual speech recognition. In: Proceedings of Conference on Language Resources and Evaluation (2004)"},{"key":"15_CR54","doi-asserted-by":"crossref","unstructured":"Fernandez-Lopez, A., Martinez, O., Sukno, F.M.: Towards estimating the upper bound of visual-speech recognition: the visual lip-reading feasibility database. In: Proceedings of Conference on Automatic Face and Gesture Recognition (2017)","DOI":"10.1109\/FG.2017.34"},{"key":"15_CR55","doi-asserted-by":"publisher","first-page":"1105","DOI":"10.1109\/TPAMI.2007.1041","volume":"29","author":"FM Sukno","year":"2007","unstructured":"Sukno, F.M., Ordas, S., Butakoff, C., Cruz, S., Frangi, A.F.: Active shape models with invariant optimal features: application to facial analysis. IEEE Trans. Pattern Anal. Mach. Intell. 29, 1105\u20131117 (2007)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR56","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60, 91\u2013110 (2004)","journal-title":"Int. J. Comput. Vis."},{"key":"15_CR57","doi-asserted-by":"publisher","first-page":"99","DOI":"10.2307\/3236261","volume":"6","author":"SB Franklin","year":"1995","unstructured":"Franklin, S.B., Gibson, D.J., Robertson, P.A., Pohlmann, J.T., Fralish, J.S.: Parallel analysis: a method for determining significant principal components. J. Veg. Sci. 6, 99\u2013106 (1995)","journal-title":"J. Veg. Sci."},{"key":"15_CR58","doi-asserted-by":"publisher","first-page":"552","DOI":"10.1109\/TSMCA.2010.2084081","volume":"41","author":"TM Khoshgoftaar","year":"2011","unstructured":"Khoshgoftaar, T.M., Van Hulse, J., Napolitano, A.: Comparing boosting and bagging techniques with noisy and imbalanced data. IEEE Trans. Syst. Man Cybern.-Part A: Syst. Hum. 41, 552\u2013568 (2011)","journal-title":"IEEE Trans. Syst. Man Cybern.-Part A: Syst. Hum."},{"key":"15_CR59","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"317","DOI":"10.1007\/3-540-44938-8_32","volume-title":"Multiple Classifier Systems","author":"S Verbaeten","year":"2003","unstructured":"Verbaeten, S., Van Assche, A.: Ensemble methods for noise elimination in classification problems. In: Windeatt, T., Roli, F. (eds.) MCS 2003. LNCS, vol. 2709, pp. 317\u2013325. Springer, Heidelberg (2003). https:\/\/doi.org\/10.1007\/3-540-44938-8_32"},{"key":"15_CR60","doi-asserted-by":"publisher","first-page":"845","DOI":"10.1109\/TNNLS.2013.2292894","volume":"25","author":"B Fr\u00e9nay","year":"2014","unstructured":"Fr\u00e9nay, B., Verleysen, M.: Classification in the presence of label noise: a survey. Neural Netw. Learn. Syst. 25, 845\u2013869 (2014)","journal-title":"Neural Netw. Learn. Syst."},{"key":"15_CR61","doi-asserted-by":"publisher","first-page":"275","DOI":"10.1007\/s10462-010-9156-z","volume":"33","author":"DF Nettleton","year":"2010","unstructured":"Nettleton, D.F., Orriols-Puig, A., Fornells, A.: A study of the effect of different types of noise on the precision of supervised learning techniques. Artif. Intell. Rev. 33, 275\u2013306 (2010)","journal-title":"Artif. Intell. Rev."},{"key":"15_CR62","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1109\/5.18626","volume":"77","author":"LR Rabiner","year":"1989","unstructured":"Rabiner, L.R.: A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 77, 257\u2013286 (1989)","journal-title":"Proc. IEEE"},{"key":"15_CR63","unstructured":"Petrushin, V.A.: Hidden Markov models: fundamentals and applications. In: Proceedings of Conference on Online Symposium for Electronics Engineer (2000)"},{"key":"15_CR64","unstructured":"Wells, J.C., et al.: SAMPA computer readable phonetic alphabet. In: Handbook of Standards and Resources for Spoken Language Systems, vol. 4 (1997)"},{"key":"15_CR65","unstructured":"Llisterri, J., Mari\u00f1o, J.B.: Spanish adaptation of SAMPA and automatic phonetic transcription. Reporte t\u00e9cnico del ESPRIT PROJECT, vol. 6819 (1993)"}],"container-title":["Communications in Computer and Information Science","Computer Vision, Imaging and Computer Graphics \u2013 Theory and Applications"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-12209-6_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,10]],"date-time":"2022-09-10T20:57:08Z","timestamp":1662843428000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-12209-6_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030122089","9783030122096"],"references-count":65,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-12209-6_15","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"23 January 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"VISIGRAPP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Joint Conference on Computer Vision, Imaging and Computer Graphics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Porto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2017","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 February 2017","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 March 2017","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"visigrapp2017","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.visigrapp.org\/?y=2017","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"Primoris","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"402","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"17","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"4% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}