{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,22]],"date-time":"2025-02-22T05:38:14Z","timestamp":1740202694930,"version":"3.37.3"},"publisher-location":"Berlin, Heidelberg","reference-count":63,"publisher":"Springer Berlin Heidelberg","isbn-type":[{"type":"print","value":"9783642126031"},{"type":"electronic","value":"9783642126048"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2010]]},"DOI":"10.1007\/978-3-642-12604-8_6","type":"book-chapter","created":{"date-parts":[[2010,6,21]],"date-time":"2010-06-21T07:37:48Z","timestamp":1277105868000},"page":"109-132","source":"Crossref","is-referenced-by-count":13,"title":["Facial Expression Synthesis Based on Emotion Dimensions for Affective Talking Avatar"],"prefix":"10.1007","author":[{"given":"Shen","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Zhiyong","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Helen M.","family":"Meng","sequence":"additional","affiliation":[]},{"given":"Lianhong","family":"Cai","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"6_CR1","doi-asserted-by":"crossref","unstructured":"Albrecht, I., Haber, J., Seidel, H.P.: Automatic generation of non-verbal facial expressions from speech. In: Proc. Computer Graphics International 2002, pp. 283\u2013293 (2002)","DOI":"10.1007\/978-1-4471-0103-1_18"},{"issue":"4","key":"6_CR2","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1007\/s10055-005-0153-5","volume":"8","author":"I. Albrecht","year":"2005","unstructured":"Albrecht, I., Schr\u00f6der, M., Haber, J., Seidel, H.P.: Mixed feelings: expression of non-basic emotions in a muscle-based talking head. Virtual Reality\u00a08(4), 201\u2013212 (2005)","journal-title":"Virtual Reality"},{"issue":"3","key":"6_CR3","doi-asserted-by":"publisher","first-page":"1075","DOI":"10.1109\/TASL.2006.885910","volume":"15","author":"C. Busso","year":"2007","unstructured":"Busso, C., Deng, Z., Grimm, M., Neumann, U., Narayanan, S.: Rigid head motion in expressive speech animation: analysis and synthesis. IEEE Transactions on Audio, Speech, and Language Processing\u00a015(3), 1075\u20131086 (2007)","journal-title":"IEEE Transactions on Audio, Speech, and Language Processing"},{"key":"6_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"450","DOI":"10.1007\/978-3-540-89646-3_44","volume-title":"Advances in Visual Computing","author":"J. Cao","year":"2008","unstructured":"Cao, J., Wang, H., Hu, P., Miao, J.: PAD model based facial expression analysis. In: Bebis, G., Boyle, R., Parvin, B., Koracin, D., Remagnino, P., Porikli, F., Peters, J., Klosowski, J., Arns, L., Chun, Y.K., Rhyne, T.-M., Monroe, L. (eds.) ISVC 2008, Part II. LNCS, vol.\u00a05359, pp. 450\u2013459. Springer, Heidelberg (2008)"},{"key":"6_CR5","doi-asserted-by":"crossref","unstructured":"Cao, Y., Tien, W.C., Faloutsos, P., Pighin, F.: Expressive speech-driven facial animation. ACM Trans. on Graph\u00a024 (2005)","DOI":"10.1145\/1095878.1095881"},{"key":"6_CR6","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1017\/S0048577299971184","volume":"36","author":"J. Cohn","year":"1999","unstructured":"Cohn, J., Zlochower, A., Lien, J., Kanade, T.: Automated face analysis by feature point tracking has high concurrent validity with manual FACS coding. Psychophysiology\u00a036, 35\u201343 (1999)","journal-title":"Psychophysiology"},{"key":"6_CR7","doi-asserted-by":"crossref","unstructured":"Cowie, R., Douglas-Cowie, E., Tsapatsoulis, N., Votsis, G., Kollias, S., Fellenz, W., Taylor, J.G.: Emotion recognition in human-computer interaction, vol.\u00a018(1), pp. 32\u201380 (2001)","DOI":"10.1109\/79.911197"},{"key":"6_CR8","unstructured":"Cowie, R., Douglas-Cowie, E., Savvidou, S., McMahon, E., Sawey, M., Schr\u00f6der, M.: FEELTRACE: an instrument for recording perceived emotion in real time. In: Proceedings of the ISCA workshop on speech and emotion, Northern Ireland, pp. 19\u201324 (2000)"},{"key":"6_CR9","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"776","DOI":"10.1007\/978-3-540-74889-2_94","volume-title":"Affective Computing and Intelligent Interaction","author":"D. Cui","year":"2007","unstructured":"Cui, D., Meng, F., Cai, L., Sun, L.: Affect related acoustic features of speech and their modification. In: Paiva, A.C.R., Prada, R., Picard, R.W. (eds.) ACII 2007. LNCS, vol.\u00a04738, pp. 776\u2013777. Springer, Heidelberg (2007)"},{"key":"6_CR10","doi-asserted-by":"crossref","DOI":"10.7208\/chicago\/9780226220802.001.0001","volume-title":"The expression of the emotions in man and animals","author":"C. Darwin","year":"1965","unstructured":"Darwin, C.: The expression of the emotions in man and animals. University of Chicago Press, Chicago (1965)"},{"issue":"16","key":"6_CR11","doi-asserted-by":"publisher","first-page":"2923","DOI":"10.1016\/S0167-8655(03)00153-3","volume":"24","author":"Y. Du","year":"2003","unstructured":"Du, Y., Lin, X.: Emotional facial expression model building. Pattern Recognition Letters\u00a024(16), 2923\u20132934 (2003)","journal-title":"Pattern Recognition Letters"},{"key":"6_CR12","unstructured":"Ekman, P.: Universals and cultural differences in facial expressions of emotion. In: Cole, J. (ed.) Proc. Nebraska Symposium on Motivation, vol.\u00a019, pp. 207\u2013283 (1971)"},{"key":"6_CR13","first-page":"169","volume-title":"Human ethology: claims and limits of a new discipline: contributions to the Colloquium","author":"P. Ekman","year":"1979","unstructured":"Ekman, P.: About brows: emotional and conversational signals. In: Human ethology: claims and limits of a new discipline: contributions to the Colloquium, pp. 169\u2013248. Cambridge University Press, England (1979)"},{"key":"6_CR14","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen, W.: Facial action coding system: A technique for the measurement of facial movement. Tech. rep. Consulting Psychologists Press (1978)","DOI":"10.1037\/t27734-000"},{"key":"6_CR15","unstructured":"Faigin, G.: The Artist\u2019s Complete Guide to Facial Expression. Watson-Guptill (2008)"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Fasel, B., Luttin, J.: Recognition of asymmetric facial action unit activities and intensities. In: Proceedings of International Conference of Pattern Recognition (2000)","DOI":"10.1109\/ICPR.2000.905664"},{"key":"6_CR17","unstructured":"Friesen, W., Ekman, P.: Emfacs-7: emotional facial action coding system, Unpublished manuscript, University of California at San Francisco (1983)"},{"key":"6_CR18","doi-asserted-by":"publisher","first-page":"238","DOI":"10.1137\/0717021","volume":"17","author":"F.N. Fritsch","year":"1980","unstructured":"Fritsch, F.N., Carlson, R.E.: Monotone piecewise cubic interpolation. SIAM Journal on Numerical Analysis\u00a017, 238\u2013246 (1980)","journal-title":"SIAM Journal on Numerical Analysis"},{"issue":"3-4","key":"6_CR19","doi-asserted-by":"publisher","first-page":"473","DOI":"10.1016\/j.specom.2005.02.017","volume":"46","author":"B. Granstrom","year":"2005","unstructured":"Granstrom, B., House, D.: Audiovisual representation of prosody in expressive speech communication. Speech Communication\u00a046(3-4), 473\u2013484 (2005)","journal-title":"Speech Communication"},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Hong, P., Wen, Z., Huang, T.S.: Real-time speech-driven face animation with expressions using neural networks, vol.\u00a013(4), pp. 916\u2013927 (2002)","DOI":"10.1109\/TNN.2002.1021892"},{"key":"6_CR21","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1007\/s10055-003-0112-y","volume":"7","author":"J. Ibanez","year":"2003","unstructured":"Ibanez, J., Aylett, R., Ruiz-Rodarte, R.: Storytelling in virtual environments from a virtual guide perspective. Virtual Reality\u00a07, 30\u201342 (2003)","journal-title":"Virtual Reality"},{"key":"6_CR22","doi-asserted-by":"crossref","unstructured":"Kalra, P., Mangili, A., Magnenat-Thalmann, N., Thalmann, D.: Simulation of facial muscle actions based on rational free form deformations. In: Proc. Eurographics 1992, pp. 59\u201369 (1992)","DOI":"10.1111\/1467-8659.1130059"},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Kanade, T., Cohn, J.F., Tian, Y.: Comprehensive database for facial expression analysis. In: Proc. Fourth IEEE International Conference on Automatic Face and Gesture Recognition, pp. 46\u201353 (2000)","DOI":"10.1109\/AFGR.2000.840611"},{"key":"6_CR24","unstructured":"Kshirsagar, S., Escher, M., Sannier, G., Magnenat-Thalmann, N.: Multimodal animation system based on the mpeg-4 standard. In: Proceedings Multimedia Modelling 1999, pp. 21\u201325 (1999)"},{"key":"6_CR25","first-page":"119","volume-title":"Technology in mental health care delivery systems","author":"P.J. Lang","year":"1980","unstructured":"Lang, P.J.: Behavioral treatment and bio-behavioral assessment: computer applications. In: Technology in mental health care delivery systems, pp. 119\u2013137. Ablex, Norwood (1980)"},{"key":"6_CR26","doi-asserted-by":"crossref","unstructured":"Lavagetto, F., Pockaj, R.: An efficient use of mpeg-4 fap interpolation for facial animation at 70 bits\/frame, vol.\u00a011(10), pp. 1085\u20131097 (2001)","DOI":"10.1109\/76.954495"},{"key":"6_CR27","volume-title":"Handbook of Facial Recognition","author":"S.Z. Li","year":"2005","unstructured":"Li, S.Z., Jain, A.K.: Handbook of Facial Recognition. Springer, New York (2005)"},{"key":"6_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"513","DOI":"10.1007\/11573548_66","volume-title":"Affective Computing and Intelligent Interaction","author":"X. Li","year":"2005","unstructured":"Li, X., Zhou, H., Song, S., Ran, T., Fu, X.: The reliability and validity of the chinese version of abbreviated PAD emotion scales. In: Tao, J., Tan, T., Picard, R.W. (eds.) ACII 2005. LNCS, vol.\u00a03784, pp. 513\u2013518. Springer, Heidelberg (2005)"},{"key":"6_CR29","unstructured":"Linden Research, Inc.: Second life: Online 3D virtual world, http:\/\/secondlife.com\/"},{"key":"6_CR30","doi-asserted-by":"crossref","unstructured":"Lyons, M., Akamatsu, S., Kamachi, M., Gyoba, J.: Coding facial expressions with gabor wavelets. In: Proc. Third IEEE International Conference on Automatic Face and Gesture Recognition, pp. 200\u2013205 (1998)","DOI":"10.1109\/AFGR.1998.670949"},{"key":"6_CR31","doi-asserted-by":"crossref","unstructured":"Mana, N., Pianesi, F.: Hmm-based synthesis of emotional facial expressions during speech in synthetic talking heads. In: Proceeding of 8th International Conference on Multimodal Interfaces (ICMI 2006), Banff, AB, Canada, pp. 380\u2013387 (2006)","DOI":"10.1145\/1180995.1181065"},{"key":"6_CR32","first-page":"53","volume":"2","author":"A. Mehrabian","year":"1968","unstructured":"Mehrabian, A.: Communication without words. Psychology Today\u00a02, 53\u201356 (1968)","journal-title":"Psychology Today"},{"issue":"4","key":"6_CR33","doi-asserted-by":"crossref","first-page":"261","DOI":"10.1007\/BF02686918","volume":"14","author":"A. Mehrabian","year":"1996","unstructured":"Mehrabian, A.: Pleasure-arousal-dominance: A general framework for describing and measuring individual differences in temperament. Current Psychology: Developmental, Learning, Personality, Social\u00a014(4), 261\u2013292 (1996)","journal-title":"Current Psychology: Developmental, Learning, Personality, Social"},{"key":"6_CR34","unstructured":"Motion Pictures Expert Group: ISO\/IEC 14496-2.: International standard, information technology-coding of audio-visual objects. part 2: Visual; amendment 1: Visual extensions (1999\/Amd. 1: 2000(E))"},{"issue":"5","key":"6_CR35","doi-asserted-by":"publisher","first-page":"843","DOI":"10.1109\/TMM.2009.2021722","volume":"11","author":"E. Mower","year":"2009","unstructured":"Mower, E., Mataric, M.J., Narayanan, S.: Human perception of audio-visual synthetic character emotion expression in the presence of ambiguous and conflicting information. IEEE Transaction on Multimedia\u00a011(5), 843\u2013855 (2009)","journal-title":"IEEE Transaction on Multimedia"},{"key":"6_CR36","unstructured":"Oddcast Inc.: Personalized speaking avatars service, http:\/\/www.voki.com\/"},{"key":"6_CR37","unstructured":"Osgood, C.E., Suci, G.J., Tannenbaum, P.H.: The Measurement of Meaning. University of Illinois Press (1957)"},{"key":"6_CR38","doi-asserted-by":"crossref","unstructured":"Parke, F.I.: Parameterized models for facial animation, vol.\u00a02(9), pp. 61\u201368 (1982)","DOI":"10.1109\/MCG.1982.1674492"},{"key":"6_CR39","doi-asserted-by":"crossref","DOI":"10.1093\/acprof:oso\/9780195179644.001.0001","volume-title":"What the face reveals: basic and applied studies of spontaneous expression using the facial action coding system (FACS)","author":"P. Ekman","year":"2005","unstructured":"Ekman, P., Rosenberg, E.L.: What the face reveals: basic and applied studies of spontaneous expression using the facial action coding system (FACS). Oxford University Press, US (2005)"},{"key":"6_CR40","doi-asserted-by":"crossref","DOI":"10.7551\/mitpress\/1140.001.0001","volume-title":"Affective Computing","author":"R.W. Picard","year":"1997","unstructured":"Picard, R.W.: Affective Computing. MIT Press, Cambridge (1997)"},{"key":"6_CR41","unstructured":"Raouzaiou, A., Tsapatsoulis, N., Karpouzis, K., Kollias, S.: Parameterized facial expression synthesis based on MPEG-4. EURASIP Journal on Applied Signal Processing 2002, 1021\u20131038 (2002)"},{"key":"6_CR42","unstructured":"Reallusion, Inc.: Crazytalk for skype, http:\/\/www.reallusion.com\/crazytalk4skype\/"},{"key":"6_CR43","series-title":"Theories of emotion","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1016\/B978-0-12-558701-3.50007-7","volume-title":"Emotion: Theory, research, and experience","author":"R. Plutchik","year":"1980","unstructured":"Plutchik, R.: A general psychoevolutionary theory of emotion. In: Emotion: Theory, research, and experience. Theories of emotion, vol.\u00a01, pp. 3\u201333. Academic, New York (1980)"},{"key":"6_CR44","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1016\/0092-6566(77)90037-X","volume":"11","author":"J. Russell","year":"1977","unstructured":"Russell, J., Mehrabian, A.: Evidence for a three-factor theory of emotions. Journal of Research in Personality\u00a011, 273\u2013294 (1977)","journal-title":"Journal of Research in Personality"},{"volume-title":"The Psychology of Facial Expression","year":"1997","key":"6_CR45","unstructured":"Russell, J.A., Fernez-Dols, J.M. (eds.): The Psychology of Facial Expression. Cambridge University Press, Cambridge (1997)"},{"issue":"1","key":"6_CR46","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1111\/1467-8659.t01-1-00645","volume":"22","author":"Z. Ruttkay","year":"2003","unstructured":"Ruttkay, Z., Noot, H., Hagen, P.: Emotion disc and emotion squares: Tools to explore the facial expression space. Computer Graphics Forum\u00a022(1), 49\u201353 (2003)","journal-title":"Computer Graphics Forum"},{"key":"6_CR47","series-title":"Lecture Notes in Artificial Intelligence","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1007\/978-3-540-24842-2_21","volume-title":"Affective Dialogue Systems","author":"M. Schr\u00f6der","year":"2004","unstructured":"Schr\u00f6der, M.: Dimensional emotion representation as a basis for speech synthesis with non-extreme emotions. In: Andr\u00e9, E., Dybkj\u00e6r, L., Minker, W., Heisterkamp, P. (eds.) ADS 2004. LNCS (LNAI), vol.\u00a03068, pp. 209\u2013220. Springer, Heidelberg (2004)"},{"issue":"4","key":"6_CR48","doi-asserted-by":"publisher","first-page":"1128","DOI":"10.1109\/TASL.2006.876118","volume":"14","author":"M. Schr\u00f6der","year":"2006","unstructured":"Schr\u00f6der, M.: Expressing degree of activation in synthetic speech. IEEE Transactions on Audio, Speech, and Language Processing\u00a014(4), 1128\u20131136 (2006)","journal-title":"IEEE Transactions on Audio, Speech, and Language Processing"},{"key":"6_CR49","doi-asserted-by":"crossref","unstructured":"Tang, H., Huang, T.S.: MPEG4 performance-driven avatar via robust facial motion tracking. In: International Conference on Image Processing, ICIP, San Diego, CA, United state, pp. 249\u2013252 (2008)","DOI":"10.1109\/ICIP.2008.4711738"},{"key":"6_CR50","doi-asserted-by":"crossref","first-page":"73","DOI":"10.1002\/vis.4340010208","volume":"1","author":"D. Terzopolous","year":"1990","unstructured":"Terzopolous, D., Waters, K.: Physically-based facial modeling, analysis and animation. Journal of Visualization and Computer Animation\u00a01, 73\u201380 (1990)","journal-title":"Journal of Visualization and Computer Animation"},{"issue":"4","key":"6_CR51","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TASL.2006.876129","volume":"14","author":"M. Theune","year":"2006","unstructured":"Theune, M., Meijs, K., Heylen, D., Ordelman, R.: Generating expressive speech for storytelling applications. IEEE Transactions on Audio, Speech, and Language Processing\u00a014(4), 1137\u20131144 (2006)","journal-title":"IEEE Transactions on Audio, Speech, and Language Processing"},{"key":"6_CR52","doi-asserted-by":"crossref","unstructured":"Tian, Y.I., Kanade, T., Cohn, J.F.: Recognizing action units for facial expression analysis, vol.\u00a023(2), pp. 97\u2013115 (2001)","DOI":"10.1109\/34.908962"},{"key":"6_CR53","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1002\/0470854626.ch9","volume-title":"MPEG-4 facial animation-the standard implementations applications","author":"N. Tsapatsoulis","year":"2002","unstructured":"Tsapatsoulis, N., Raousaiou, A., Kollias, S., Cowie, R., Douglas-Cowie, E.: Emotion recognition and synthesis based on MPEG-4 FAPs. In: MPEG-4 facial animation-the standard implementations applications, pp. 141\u2013167. Wiley, Hillsdale (2002)"},{"key":"6_CR54","doi-asserted-by":"crossref","unstructured":"Wang, Z., Cai, L., AI, H.: A dynamic viseme model for personalizing a talking head. In: Sixth International Conference on Signal Processing (ICSP 2002), pp. 26\u201330 (2002)","DOI":"10.1109\/ICOSP.2002.1179960"},{"issue":"4","key":"6_CR55","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1145\/37402.37405","volume":"22","author":"K. Waters","year":"1987","unstructured":"Waters, K.: A muscle model of animating three dimensional facial expression. Computer Graphics\u00a022(4), 17\u201324 (1987)","journal-title":"Computer Graphics"},{"key":"6_CR56","series-title":"Lecture Notes in Artificial Intelligence","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1007\/11590323_21","volume-title":"Intelligent Technologies for Interactive Entertainment","author":"H. Welbergen","year":"2005","unstructured":"Welbergen, H., Nijholt, A., Reidsma, D., Zwiers, J.: Presenting in virtual worlds: Towards an architecture for a 3D presenter explaining 2D-presented information. In: Maybury, M., Stock, O., Wahlster, W. (eds.) INTETAIN 2005. LNCS (LNAI), vol.\u00a03814, pp. 203\u2013212. Springer, Heidelberg (2005)"},{"key":"6_CR57","doi-asserted-by":"crossref","first-page":"113","DOI":"10.1016\/B978-0-12-558704-4.50011-6","volume-title":"The Measurement of Emotions","author":"C. Whissell","year":"1989","unstructured":"Whissell, C.: The Dictionary of Affect in Language Emotion: Theory, Research and Experience. In: The Measurement of Emotions, vol.\u00a04, pp. 113\u2013131. Academic Press, London (1989)"},{"key":"6_CR58","doi-asserted-by":"crossref","unstructured":"Wu, Z., Zhang, S., Cai, L., Meng, H.M.: Real-time synthesis of Chinese visual speech and facial expressions using mpeg-4 fap features in a three-dimensional avatar. In: INTERSPEECH 2006 and 9th International Conference on Spoken Language Processing, vol.\u00a04, pp. 1802\u20131805 (2006)","DOI":"10.21437\/Interspeech.2006-498"},{"key":"6_CR59","doi-asserted-by":"crossref","unstructured":"Yang, H., Meng, H.M., Cai, L.: Modeling the acoustic correlates of expressive elements in text genres for expressive text-to-speech synthesis. In: INTERSPEECH 2006 and 9th International Conference on Spoken Language Processing, vol.\u00a04, pp. 1806\u20131809 (2006)","DOI":"10.21437\/Interspeech.2006-499"},{"key":"6_CR60","doi-asserted-by":"crossref","unstructured":"Yang, H., Meng, H.M., Wu, Z., Cai, L.: Modelling the global acoustic correlates of expressivity for Chinese text-to-speech synthesis. In: Proc. IEEE Spoken Language Technology Workshop, pp. 138\u2013141 (2006)","DOI":"10.1109\/SLT.2006.326837"},{"issue":"1","key":"6_CR61","first-page":"39","volume":"31","author":"Z. Zeng","year":"2009","unstructured":"Zeng, Z., Pantic, M., Roisman, G.I., Huang, T.S.: A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Transaction on Multimedia\u00a031(1), 39\u201358 (2009)","journal-title":"IEEE Transaction on Multimedia"},{"key":"6_CR62","unstructured":"Zhang, S.: Pseudo facial expression database, http:\/\/hcsi.cs.tsinghua.edu.cn\/Demo\/jaffe\/emot\/index.php"},{"key":"6_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, S., Wu, Z., Meng, H.M., Cai, L.: Head movement synthesis based on semantic and prosodic features for a Chinese expressive avatar. In: Proc. IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP 2007, vol.\u00a04, pp. 837\u2013840 (2007)","DOI":"10.1109\/ICASSP.2007.367043"}],"container-title":["Smart Innovation, Systems and Technologies","Modeling Machine Emotions for Realizing Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-642-12604-8_6.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,22]],"date-time":"2025-02-22T03:06:22Z","timestamp":1740193582000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-642-12604-8_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2010]]},"ISBN":["9783642126031","9783642126048"],"references-count":63,"URL":"https:\/\/doi.org\/10.1007\/978-3-642-12604-8_6","relation":{},"ISSN":["2190-3018","2190-3026"],"issn-type":[{"type":"print","value":"2190-3018"},{"type":"electronic","value":"2190-3026"}],"subject":[],"published":{"date-parts":[[2010]]}}}