{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2022,3,30]],"date-time":"2022-03-30T14:56:55Z","timestamp":1648652215833},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2012,9,13]],"date-time":"2012-09-13T00:00:00Z","timestamp":1347494400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2013,10]]},"DOI":"10.1007\/s00371-012-0751-7","type":"journal-article","created":{"date-parts":[[2012,9,12]],"date-time":"2012-09-12T08:56:25Z","timestamp":1347440185000},"page":"969-982","source":"Crossref","is-referenced-by-count":6,"title":["Automatic visual speech segmentation and recognition using directional motion history images and Zernike moments"],"prefix":"10.1007","volume":"29","author":[{"given":"Ayaz A.","family":"Shaikh","sequence":"first","affiliation":[]},{"given":"Dinesh K.","family":"Kumar","sequence":"additional","affiliation":[]},{"given":"Jayavardhana","family":"Gubbi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2012,9,13]]},"reference":[{"issue":"6","key":"751_CR1","doi-asserted-by":"crossref","first-page":"1161","DOI":"10.1109\/TSA.2005.853002","volume":"13","author":"C. Xiaodong","year":"2005","unstructured":"Xiaodong, C., Alwan, A.: Noise robust speech recognition using feature compensation based on polynomial regression of utterance SNR. IEEE Trans. Speech Audio Process. 13(6), 1161\u20131172 (2005)","journal-title":"IEEE Trans. Speech Audio Process."},{"key":"751_CR2","doi-asserted-by":"crossref","first-page":"701","DOI":"10.1109\/LSP.2008.2001820","volume":"15","author":"X. Haitian","year":"2008","unstructured":"Haitian, X., Zheng-Hua, T., Dalsgaard, P., Lindberg, B.: Robust speech recognition by nonlocal means denoising processing. IEEE Signal Process. Lett. 15, 701\u2013704 (2008)","journal-title":"IEEE Signal Process. Lett."},{"issue":"5","key":"751_CR3","doi-asserted-by":"crossref","first-page":"798","DOI":"10.1109\/JSTSP.2010.2057192","volume":"4","author":"T. Zheng-Hua","year":"2010","unstructured":"Zheng-Hua, T., Lindberg, B.: Low-complexity variable frame rate analysis for speech recognition and voice activity detection. IEEE J. Sel. Top. Signal Process. 4(5), 798\u2013807 (2010)","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"751_CR4","first-page":"265","volume-title":"IEEE Global Telecommunications Conference","author":"E. Petajan","year":"1984","unstructured":"Petajan, E.: Automatic lipreading to enhance speech recognition. In: IEEE Global Telecommunications Conference, Atlanta, GA, USA, pp. 265\u2013272. IEEE Computer Society Press, Los Alamitos (1984)"},{"key":"751_CR5","doi-asserted-by":"crossref","first-page":"2191","DOI":"10.1109\/IEMBS.2006.260213","volume-title":"Engineering in Medicine and Biology Society. EMBS\u201906. 28th Annual International Conference of the IEEE","author":"S.P. Arjunan","year":"2006","unstructured":"Arjunan, S.P., Kumar, D.K., Yau, W.C., Weghorn, H.: Unspoken vowel recognition using facial electromyogram. In: Engineering in Medicine and Biology Society. EMBS\u201906. 28th Annual International Conference of the IEEE, Aug. 30 2006\u2013Sept.\u00a03 2006, pp. 2191\u20132194 (2006)"},{"issue":"4","key":"751_CR6","doi-asserted-by":"crossref","first-page":"341","DOI":"10.1016\/j.specom.2009.12.002","volume":"52","author":"T. Schultz","year":"2010","unstructured":"Schultz, T., Wand, M.: Modeling coarticulation in EMG-based continuous speech recognition. Speech Commun. 52(4), 341\u2013353 (2010). doi: 10.1016\/j.specom.2009.12.002","journal-title":"Speech Commun."},{"key":"751_CR7","unstructured":"Medizinelektronik, C.: (2008). http:\/\/www.articulograph.de\/"},{"key":"751_CR8","unstructured":"Soquet, A., Saerens, M., Lecuit, V.: Complementary cues for speech recognition, pp.\u00a01645\u20131648 (1999)"},{"issue":"9","key":"751_CR9","doi-asserted-by":"crossref","first-page":"1306","DOI":"10.1109\/JPROC.2003.817150","volume":"91","author":"G. Potamianos","year":"2003","unstructured":"Potamianos, G., Neti, C., Gravier, G., Garg, A., Senior, A.W.: Recent advances in the automatic recognition of audiovisual speech. Proc. IEEE 91(9), 1306\u20131326 (2003)","journal-title":"Proc. IEEE"},{"issue":"3","key":"751_CR10","doi-asserted-by":"crossref","first-page":"419","DOI":"10.1142\/S0219467808003167","volume":"8","author":"W.C. Yau","year":"2008","unstructured":"Yau, W.C., Kumar, D.K., Arjunan, S.P.: Visual speech recognition using dynamic features and support vector machines. Int. J. Image Graph. 8(3), 419\u2013437 (2008)","journal-title":"Int. J. Image Graph."},{"issue":"1","key":"751_CR11","doi-asserted-by":"crossref","first-page":"21","DOI":"10.1007\/s11263-006-4329-6","volume":"67","author":"T. Xiang","year":"2006","unstructured":"Xiang, T., Gong, S.: Beyond tracking: modelling activity and understanding behaviour. Int. J. Comput. Vis. 67(1), 21\u201351 (2006)","journal-title":"Int. J. Comput. Vis."},{"key":"751_CR12","first-page":"21","volume-title":"Proc. Computer Vision Theory and Applications","author":"H. Meng","year":"2007","unstructured":"Meng, H., Pears, N., Bailey, C.: Motion information combination for fast human action recognition. In: Proc. Computer Vision Theory and Applications, Spain, pp. 21\u201328 (2007)"},{"issue":"5","key":"751_CR13","doi-asserted-by":"crossref","first-page":"485","DOI":"10.1002\/cav.11","volume":"15","author":"J. Ma","year":"2004","unstructured":"Ma, J., Cole, R., Pellom, B., Ward, W., Wise, B.: Accurate automatic visible speech synthesis of arbitrary 3D models based on concatenation of diviseme motion capture data. Comput. Animat. Virtual Worlds 15(5), 485\u2013500 (2004)","journal-title":"Comput. Animat. Virtual Worlds"},{"key":"751_CR14","unstructured":"Govokhina, O., Bailly, G., Breton, G.: Learning Optimal Audiovisual Phasing for a HMM-based Control Model for Facial Animation (2007)"},{"key":"751_CR15","doi-asserted-by":"crossref","first-page":"1401","DOI":"10.21437\/Interspeech.2010-426","volume-title":"Proc. of the Interspeech 2010","author":"U. Musti","year":"2010","unstructured":"Musti, U., Toutios, A., Ouni, S., Colotte, V., Wrobel-Dautcourt, B., Berger, M.O.: HMM-based automatic visual speech segmentation using facial data. In: Proc. of the Interspeech 2010, pp. 1401\u20131404 (2010)"},{"issue":"5","key":"751_CR16","doi-asserted-by":"crossref","first-page":"477","DOI":"10.1016\/S0923-5965(00)00011-4","volume":"16","author":"I. Koprinska","year":"2001","unstructured":"Koprinska, I., Carrato, S.: Temporal video segmentation: a survey. Signal Process. Image Commun. 16(5), 477\u2013500 (2001)","journal-title":"Signal Process. Image Commun."},{"key":"751_CR17","doi-asserted-by":"crossref","first-page":"399","DOI":"10.1109\/SIBGRA.2003.1241036","volume-title":"XVI Brazilian Symposium on Computer Graphics and Image Processing SIBGRAPI 2003","author":"L.G. Silveira Da","year":"2003","unstructured":"Da Silveira, L.G., Facon, J., Borges, D.L.: Visual speech recognition: a solution from feature extraction to words classification. In: XVI Brazilian Symposium on Computer Graphics and Image Processing SIBGRAPI 2003, pp. 399\u2013405 (2003)"},{"key":"751_CR18","doi-asserted-by":"crossref","first-page":"383","DOI":"10.1007\/978-3-662-13015-5_28","volume":"150","author":"J. Luettin","year":"1996","unstructured":"Luettin, J., Thacker, N.A., Beet, S.W.: Active shape models for visual speech feature extraction. NATO ASI Ser. Comput. Syst. Sci. 150, 383\u2013390 (1996)","journal-title":"NATO ASI Ser. Comput. Syst. Sci."},{"issue":"1","key":"751_CR19","doi-asserted-by":"crossref","first-page":"42","DOI":"10.1109\/49.363147","volume":"13","author":"K. Otani","year":"1995","unstructured":"Otani, K., Hasegawa, T.: The image input microphone\u2014a new nonacoustic speech communication system by media conversion from oral motion images to speech. IEEE J. Sel. Areas Commun. 13(1), 42\u201348 (1995)","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"751_CR20","first-page":"2193","volume-title":"16th International Congress of Phonetic Sciences","author":"T. Hueber","year":"2007","unstructured":"Hueber, T., Chollet, G., Denby, B., Stone, M., Zouari, L.: Ouisper: corpus based synthesis driven by articulatory data. In: 16th International Congress of Phonetic Sciences, pp. 2193\u20132196 (2007)"},{"issue":"11","key":"751_CR21","doi-asserted-by":"crossref","first-page":"65","DOI":"10.1109\/35.41402","volume":"27","author":"B.P. Yuhas","year":"1989","unstructured":"Yuhas, B.P., Goldstein, M.H. Jr., Sejnowski, T.J.: Integration of acoustic and visual speech signals using neural networks. IEEE Commun. Mag. 27(11), 65\u201371 (1989)","journal-title":"IEEE Commun. Mag."},{"key":"751_CR22","first-page":"669","volume-title":"IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP-94","author":"C. Bregler","year":"1994","unstructured":"Bregler, C., Konig, Y.: Eigenlips for robust speech recognition. In: IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP-94, 19\u201322 Apr 1994, pp.\u00a0669\u2013672 (1994)"},{"issue":"5","key":"751_CR23","doi-asserted-by":"crossref","first-page":"337","DOI":"10.1109\/89.536928","volume":"4","author":"P.L. Silsbee","year":"1996","unstructured":"Silsbee, P.L., Bovik, A.C.: Computer lipreading for improved accuracy in automatic speech recognition. IEEE Trans. Speech Audio Process. 4(5), 337\u2013351 (1996)","journal-title":"IEEE Trans. Speech Audio Process."},{"key":"751_CR24","first-page":"165","volume-title":"Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP\u201901)","author":"G. Potamianos","year":"2001","unstructured":"Potamianos, G., Luettin, J., Neti, C.: Hierarchical discriminant features for audio-visual LVCSR. In: Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP\u201901), vol.\u00a0161, pp. 165\u2013168 (2001)"},{"issue":"7","key":"751_CR25","doi-asserted-by":"crossref","first-page":"1254","DOI":"10.1109\/TMM.2009.2030637","volume":"11","author":"G. Zhao","year":"2009","unstructured":"Zhao, G., Barnard, M., Pietikainen, M.: Lipreading with local spatiotemporal descriptors. IEEE Trans. Multimed. 11(7), 1254\u20131265 (2009)","journal-title":"IEEE Trans. Multimed."},{"issue":"2","key":"751_CR26","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1007\/BF00127169","volume":"8","author":"A.L. Yuille","year":"1992","unstructured":"Yuille, A.L., Hallinan, P.W., Cohen, D.S.: Feature extraction from faces using deformable templates. Int. J. Comput. Vis. 8(2), 99\u2013111 (1992)","journal-title":"Int. J. Comput. Vis."},{"issue":"4","key":"751_CR27","doi-asserted-by":"crossref","first-page":"321","DOI":"10.1007\/BF00133570","volume":"1","author":"M. Kass","year":"1988","unstructured":"Kass, M., Witkin, A., Terzopoulos, D.: Snakes: active contour models. Int. J. Comput. Vis. 1(4), 321\u2013331 (1988)","journal-title":"Int. J. Comput. Vis."},{"issue":"3","key":"751_CR28","doi-asserted-by":"crossref","first-page":"423","DOI":"10.1109\/TASL.2008.2011515","volume":"17","author":"G. Papandreou","year":"2009","unstructured":"Papandreou, G., Katsamanis, A., Pitsikalis, V., Maragos, P.: Adaptive multimodal fusion by uncertainty compensation with application to audiovisual speech recognition. IEEE Trans. Audio Speech Lang. Process. 17(3), 423\u2013435 (2009)","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"issue":"2","key":"751_CR29","doi-asserted-by":"crossref","first-page":"198","DOI":"10.1109\/34.982900","volume":"24","author":"I. Matthews","year":"2002","unstructured":"Matthews, I., Cootes, T., Bangham, J., Cox, S., Harvey, R.: Extraction of visual features for lipreading. IEEE Trans. Pattern Anal. Mach. Intell. 24(2), 198\u2013213 (2002)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"751_CR30","first-page":"572","volume-title":"Conference Record of the Twenty-Eighth Asilomar Conference on Signals, Systems and Computers","author":"A.J. Goldschen","year":"1994","unstructured":"Goldschen, A.J., Garcia, O.N., Petajan, E.: Continuous optical automatic speech recognition by lipreading. In: Conference Record of the Twenty-Eighth Asilomar Conference on Signals, Systems and Computers, 31 Oct\u20132 Nov, pp. 572\u2013577 (1994)"},{"issue":"6","key":"751_CR31","doi-asserted-by":"crossref","first-page":"67","DOI":"10.1002\/scj.4690220607","volume":"22","author":"K. Mase","year":"1991","unstructured":"Mase, K., Pentland, A.: Automatic lipreading by optical-flow analysis. Syst. Comput. Jpn. 22(6), 67\u201376 (1991)","journal-title":"Syst. Comput. Jpn."},{"issue":"1","key":"751_CR32","first-page":"4","volume":"2007","author":"K. Iwano","year":"2007","unstructured":"Iwano, K., Yoshinaga, T., Tamura, S., Furui, S.: Audio-visual speech recognition using lip information extracted from side-face images. EURASIP J. Audio Speech Music Process. 2007(1), 4 (2007)","journal-title":"EURASIP J. Audio Speech Music Process."},{"issue":"1","key":"751_CR33","first-page":"23","volume":"4","author":"R. Rajavel","year":"2010","unstructured":"Rajavel, R., Sathidevi, P.S.: A novel algorithm for acoustic and visual classifiers decision fusion in audio-visual speech recognition system. Signal Process. 4(1), 23\u201337 (2010)","journal-title":"Signal Process."},{"issue":"8","key":"751_CR34","doi-asserted-by":"crossref","first-page":"597","DOI":"10.1016\/j.imavis.2003.11.004","volume":"22","author":"R. Venkatesh Babu","year":"2004","unstructured":"Venkatesh Babu, R., Ramakrishnan, K.: Recognition of human actions using motion history information extracted from the compressed video. Image Vis. Comput. 22(8), 597\u2013607 (2004)","journal-title":"Image Vis. Comput."},{"issue":"2\u20133","key":"751_CR35","doi-asserted-by":"crossref","first-page":"249","DOI":"10.1016\/j.cviu.2006.07.013","volume":"104","author":"D. Weinland","year":"2006","unstructured":"Weinland, D., Ronfard, R., Boyer, E.: Free viewpoint action recognition using motion history volumes. Comput. Vis. Image Underst. 104(2\u20133), 249\u2013257 (2006)","journal-title":"Comput. Vis. Image Underst."},{"key":"751_CR36","first-page":"253","volume-title":"13th IEEE International Workshop on Robot and Human Interactive Communication, ROMAN","author":"M. Valstar","year":"2004","unstructured":"Valstar, M., Patras, I., Pantic, M.: Facial action unit recognition using temporal templates. In: 13th IEEE International Workshop on Robot and Human Interactive Communication, ROMAN, 20\u201322 Sept 2004, pp. 253\u2013258 (2004)"},{"issue":"1","key":"751_CR37","doi-asserted-by":"crossref","first-page":"36","DOI":"10.4304\/jmm.5.1.36-46","volume":"5","author":"M. Ahad","year":"2010","unstructured":"Ahad, M.: Analysis of motion self-occlusion problem due to motion overwriting for human activity recognition. J. Multimed. 5(1), 36\u201346 (2010)","journal-title":"J. Multimed."},{"key":"751_CR38","volume-title":"The HTK Book","author":"S. Young","year":"1997","unstructured":"Young, S., Evermann, G., Kershaw, D., Moore, G., Odell, J., Ollason, D., Valtchev, V., Woodland, P.: The HTK Book, vol.\u00a02. Entropic Cambridge Research Laboratory, Cambridge (1997)"},{"key":"751_CR39","first-page":"42","volume-title":"Proc. International Conference on Spoken Language Processing","author":"Q. Su","year":"2002","unstructured":"Su, Q., Silsbee, P.L.: Robust audiovisual integration using semicontinuous hidden Markov models. In: Proc. International Conference on Spoken Language Processing, Philadelphia, PA, pp. 42\u201345 (2002)"},{"key":"751_CR40","first-page":"57","volume-title":"Proc. European Tutorial Workshop on Audio-Visual Speech Processing","author":"G. Krone","year":"1997","unstructured":"Krone, G., Talk, B., Wichert, A., Palm, G.: Neural architectures for sensor fusion in speech recognition. In: Proc. European Tutorial Workshop on Audio-Visual Speech Processing, Rhodes, Greece, pp. 57\u201360 (1997)"},{"key":"751_CR41","first-page":"93","volume-title":"Proc. of HCSNet Workshop on the Use of Vision in HCI","author":"W. Yau","year":"2006","unstructured":"Yau, W., Kumar, D., Arjunan, S.: Voiceless speech recognition using dynamic visual speech features. In: Proc. of HCSNet Workshop on the Use of Vision in HCI, Canberra, Australia, pp. 93\u2013101. Australian Computer Society, Inc., Canberra (2006)"},{"key":"751_CR42","first-page":"547","volume-title":"Proceedings of the International Conference on Spoken Language and Processing","author":"P. Duchnowski","year":"1994","unstructured":"Duchnowski, P., Meier, U., Waibel, A.: See me, hear me: integrating automatic speech recognition and lip-reading. In: Proceedings of the International Conference on Spoken Language and Processing, Yokohama, Japan, pp. 547\u2013550. Citeseer, Princeton (1994)"},{"key":"751_CR43","first-page":"189","volume-title":"International Conference on Auditory-Visual Speech Processing","author":"M. Heckmann","year":"2001","unstructured":"Heckmann, M., Berthommier, F., Kroschel, K.: A hybrid ANN\/HMM audio-visual speech recognition system. In: International Conference on Auditory-Visual Speech Processing, Aalborg, Denmark, pp. 189\u2013194. Citeseer, Princeton (2001)"},{"key":"751_CR44","doi-asserted-by":"crossref","first-page":"610","DOI":"10.1007\/978-3-540-85920-8_74","volume-title":"Progress in Pattern Recognition, Image Analysis and Applications","author":"W. Yau","year":"2008","unstructured":"Yau, W., Kant Kumar, D., Chinnadurai, T.: Lip-reading technique using spatio-temporal templates and support vector machines. In: Progress in Pattern Recognition, Image Analysis and Applications, pp. 610\u2013617 (2008)"},{"key":"751_CR45","first-page":"504","volume-title":"International Conference on Spoken Language Processing","author":"A. Ganapathiraju","year":"2000","unstructured":"Ganapathiraju, A., Hamaker, J., Picone, J.: Hybrid SVM\/HMM architectures for speech recognition. In: International Conference on Spoken Language Processing, pp. 504\u2013507. Citeseer, Princeton (2000)"},{"issue":"1","key":"751_CR46","doi-asserted-by":"crossref","first-page":"1248","DOI":"10.1155\/S1110865702207039","volume":"2002","author":"M. Gordan","year":"2002","unstructured":"Gordan, M., Kotropoulos, C., Pitas, I.: A support vector machine-based dynamic network for visual speech recognition applications. EURASIP J. Appl. Signal Process. 2002(1), 1248\u20131259 (2002)","journal-title":"EURASIP J. Appl. Signal Process."},{"key":"751_CR47","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"crossref","first-page":"83","DOI":"10.1007\/978-3-540-88690-7_7","volume-title":"Computer Vision\u2014ECCV 2008","author":"D. Sun","year":"2008","unstructured":"Sun, D., Roth, S., Lewis, J., Black, M.: Learning optical flow. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) Computer Vision\u2014ECCV 2008. Lecture Notes in Computer Science, vol. 5304, pp. 83\u201397. Springer, Berlin (2008)"},{"issue":"1","key":"751_CR48","doi-asserted-by":"crossref","first-page":"62","DOI":"10.1109\/TSMC.1979.4310076","volume":"9","author":"N. Otsu","year":"1979","unstructured":"Otsu, N.: A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 9(1), 62\u201366 (1979)","journal-title":"IEEE Trans. Syst. Man Cybern."},{"issue":"5","key":"751_CR49","doi-asserted-by":"crossref","first-page":"489","DOI":"10.1109\/34.55109","volume":"12","author":"A. Khotanzad","year":"1990","unstructured":"Khotanzad, A., Hong, Y.H.: Invariant image recognition by Zernike moments. IEEE Trans. Pattern Anal. Mach. Intell. 12(5), 489\u2013497 (1990)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"4","key":"751_CR50","doi-asserted-by":"crossref","first-page":"496","DOI":"10.1109\/34.3913","volume":"10","author":"C.H. Teh","year":"1988","unstructured":"Teh, C.H., Chin, R.T.: On image analysis by the methods of moments. IEEE Trans. Pattern Anal. Mach. Intell. 10(4), 496\u2013513 (1988)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"751_CR51","unstructured":"Chang, C.C., Lin, C.J.: LIBSVM: a Library for Support Vector Machines (2001)"},{"key":"751_CR52","volume-title":"Video Analysis of Mouth Movement Using Motion Templates for Computer-Based Lip-Reading","author":"W.C. Yau","year":"2008","unstructured":"Yau, W.C.: Video Analysis of Mouth Movement Using Motion Templates for Computer-Based Lip-Reading. RMIT University, Melbourne (2008)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-012-0751-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s00371-012-0751-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-012-0751-7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,28]],"date-time":"2022-01-28T19:26:58Z","timestamp":1643398018000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s00371-012-0751-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2012,9,13]]},"references-count":52,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2013,10]]}},"alternative-id":["751"],"URL":"https:\/\/doi.org\/10.1007\/s00371-012-0751-7","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2012,9,13]]}}}