{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T03:29:32Z","timestamp":1764646172186,"version":"3.41.0"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783319893266"},{"type":"electronic","value":"9783319893273"}],"license":[{"start":{"date-parts":[[2018,6,2]],"date-time":"2018-06-02T00:00:00Z","timestamp":1527897600000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-319-89327-3_14","type":"book-chapter","created":{"date-parts":[[2018,6,1]],"date-time":"2018-06-01T16:25:45Z","timestamp":1527870345000},"page":"181-196","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Multi-modal Intention Prediction with Probabilistic Movement Primitives"],"prefix":"10.1007","author":[{"given":"Oriane","family":"Dermy","sequence":"first","affiliation":[]},{"given":"Francois","family":"Charpillet","sequence":"additional","affiliation":[]},{"given":"Serena","family":"Ivaldi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,6,2]]},"reference":[{"key":"14_CR1","unstructured":"Anzalone, S.M., Boucenna, S., Ivaldi, S., Chetouani, M.: Evaluating the engagement with social robots. Int. J. Soc. Robot. 7(4), 465\u2013478 (2015)"},{"key":"14_CR2","doi-asserted-by":"crossref","unstructured":"Bader, T., Vogelgesang, M., Klaus, E.: Multimodal integration of natural gaze behavior for intention recognition during object manipulation. In: Proceedings of PIC on Multimodal Interfaces, pp. 199\u2013206. ACM (2009)","DOI":"10.1145\/1647314.1647350"},{"key":"14_CR3","unstructured":"Baluja, S., Pomerleau, D.: Non-intrusive gaze tracking using artificial neural networks. In: Proceedings of Advances in NIPS, pp. 753\u2013760 (1994)"},{"key":"14_CR4","unstructured":"Boucenna, S., Gaussier, P., Andry, P., Hafemeister, L.: A robot learns the facial expressions recognition and face\/non-face discrimination through an imitation game. Int. J. Soc. Robot. 6(4), 633\u2013652 (2014)"},{"key":"14_CR5","unstructured":"Bretherton, I.: Intentional communication and the development of an understanding of mind. Children\u2019s Theories of Mind: Mental States and Social Understanding, pp. 49\u201375 (1991)"},{"key":"14_CR6","doi-asserted-by":"crossref","unstructured":"Castellano, G., Pereira, A., Leite, I., Paiva, A., McOwan, P.W.: Detecting user engagement with a robot companion using task and social interaction-based features. In: Proceedings of PIC on Multimodal Interfaces, pp. 119\u2013126. ACM (2009)","DOI":"10.1145\/1647314.1647336"},{"key":"14_CR7","doi-asserted-by":"crossref","unstructured":"Dermy, O., Paraschos, A., Ewerton, M., Peters, J., Charpillet, F., Ivaldi, S.: Prediction of intention during interaction with ICUB with probabilistic movement primitives. Front. Robot AI (2017)","DOI":"10.3389\/frobt.2017.00045"},{"key":"14_CR8","doi-asserted-by":"crossref","unstructured":"Dillmann, R., Becher, R., Steinhaus, P.: ARMAR II-a learning and cooperative multimodal humanoid robot system. Int. J. Humanoid Robot 1(01), 143\u2013155 (2004)","DOI":"10.1142\/S0219843604000046"},{"key":"14_CR9","doi-asserted-by":"crossref","unstructured":"Dragan, A., Srinivasa, S.: Generating legible motion. In: Proceedings of Robotics: Science and Systems. Berlin, Germany, June 2013","DOI":"10.15607\/RSS.2013.IX.024"},{"key":"14_CR10","unstructured":"Dragan, A., Srinivasa, S.: Integrating human observer inferences into robot motion planning. Auton. Robot. 37(4), 351\u2013368 (2014)"},{"key":"14_CR11","unstructured":"Ferrer, G., Sanfeliu, A.: Bayesian human motion intentionality prediction in urban environments. Pattern Recogn. Lett. 44, 134\u2013140 (2014)"},{"key":"14_CR12","doi-asserted-by":"crossref","unstructured":"Hoffman, M.W., Grimes, D.B., Shon, A.P., Rao, R.P.: A probabilistic model of gaze imitation and shared attention. Neural Netw. 19(3), 299\u2013310 (2006)","DOI":"10.1016\/j.neunet.2006.02.008"},{"key":"14_CR13","doi-asserted-by":"crossref","unstructured":"Huang, C.M., Mutlu, B.: Anticipatory robot control for efficient human-robot collaboration. In: Proceedings of HRI, pp. 83\u201390 (2016)","DOI":"10.1109\/HRI.2016.7451737"},{"key":"14_CR14","unstructured":"Ishii, R., Shinohara, Y., Nakano, T., Nishida, T.: Combining multiple types of eye-gaze information to predict user\u2019s conversational engagement. In: 2nd Workshop on Eye Gaze on Intelligent Human Machine Interaction (2011)"},{"key":"14_CR15","unstructured":"Ivaldi, S., Lefort, S., Peters, J., Chetouani, M., Provasi, J., Zibetti, E.: Towards engagement models that consider individual factors in HRI. Int. J. of Soc. Robot. 9, 63\u201386 (2017)"},{"key":"14_CR16","doi-asserted-by":"crossref","unstructured":"Kim, J., Banks, C.J., Shah, J.A.: Collaborative planning with encoding of users\u2019 high-level strategies. In: Proceedings of AAAI (2017)","DOI":"10.1609\/aaai.v31i1.10627"},{"key":"14_CR17","unstructured":"Kozima, H., Yano, H.: A robot that learns to communicate with human caregivers. In: Proceedings of the First International Workshop on Epigenetic Robotics, pp. 47\u201352 (2001)"},{"key":"14_CR18","unstructured":"Ma, C., Prendinger, H., Ishizuka, M.: Eye movement as an indicator of users\u2019 involvement with embodied interfaces at the low level. In: Proceedings of AISB, pp. 136\u2013143 (2005)"},{"key":"14_CR19","unstructured":"Meltzoff, A.N., Brooks, R.: Eyes wide shut: the importance of eyes in infant gaze following and understanding other minds. In: Flom, R., Lee, K., Muir, D. (eds.) Gaze Following: Its Development and Significance. Erlbaum. [EVH] (2007)"},{"key":"14_CR20","doi-asserted-by":"crossref","unstructured":"Mitsugami, I., Ukita, N., Kidode, M.: Robot navigation by eye pointing. Lecture notes in computer science 3711, 256 (2005)","DOI":"10.1007\/11558651_26"},{"key":"14_CR21","unstructured":"Paraschos, A., Daniel, C., Peters, J.R., Neumann, G.: Probabilistic movement primitives. In: Proceedings of NIPS, pp. 2616\u20132624 (2013)"},{"key":"14_CR22","doi-asserted-by":"crossref","unstructured":"Ravichandar, H., Kumar, A., Dani, A.: Bayesian human intention inference through multiple model filtering with gaze-based priors. In: Proceedings of Information Fusion (FUSION), pp. 2296\u20132302. IEEE (2016)","DOI":"10.1109\/MFI.2015.7295812"},{"key":"14_CR23","doi-asserted-by":"crossref","unstructured":"Timm, F., Barth, E.: Accurate eye centre localisation by means of gradients. In: Proceedings of Visapp, vol. 11, pp. 125\u2013130 (2011)","DOI":"10.5220\/0003326101250130"},{"key":"14_CR24","doi-asserted-by":"crossref","unstructured":"Traver, V.J., del Pobil, A.P., P\u00e9rez-Francisco, M.: Making service robots human-safe. In: Proceedings of (IROS 2000), vol.\u00a01, pp. 696\u2013701. IEEE (2000)","DOI":"10.1109\/IROS.2000.894685"},{"key":"14_CR25","doi-asserted-by":"crossref","unstructured":"Walker-Andrews, A.S.: Infants\u2019 perception of expressive behaviors: differentiation of multimodal information. Psychol. Bull. 121(3), 437 (1997)","DOI":"10.1037\/0033-2909.121.3.437"},{"key":"14_CR26","doi-asserted-by":"crossref","unstructured":"Wang, Z., Deisenroth, M.P., Amor, H.B., Vogt, D., Sch\u00f6lkopf, B., Peters, J.: Probabilistic modeling of human movements for intention inference. Robot. Sci. Syst. (2012)","DOI":"10.15607\/RSS.2012.VIII.055"},{"key":"14_CR27","doi-asserted-by":"crossref","unstructured":"Weser, M., Westhoff, D., Huser, M., Zhang, J.: Multimodal people tracking and trajectory prediction based on learned generalized motion patterns. In: International Conference on Multisensor Fusion and Integration for Intelligent Systems, pp. 541\u2013546 (2006)","DOI":"10.1109\/MFI.2006.265639"},{"key":"14_CR28","doi-asserted-by":"crossref","unstructured":"Xiong, X., De la Torre, F.: Supervised descent method and its applications to face alignment. In: Proceedings of IEEE CVPR (2013)","DOI":"10.1109\/CVPR.2013.75"}],"container-title":["Springer Proceedings in Advanced Robotics","Human Friendly Robotics"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-319-89327-3_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,4]],"date-time":"2025-07-04T23:05:31Z","timestamp":1751670331000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-319-89327-3_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,6,2]]},"ISBN":["9783319893266","9783319893273"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-319-89327-3_14","relation":{},"ISSN":["2511-1256","2511-1264"],"issn-type":[{"type":"print","value":"2511-1256"},{"type":"electronic","value":"2511-1264"}],"subject":[],"published":{"date-parts":[[2018,6,2]]}}}