{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T08:57:06Z","timestamp":1765357026267,"version":"3.40.3"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030891336"},{"type":"electronic","value":"9783030891343"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-89134-3_27","type":"book-chapter","created":{"date-parts":[[2021,10,17]],"date-time":"2021-10-17T10:39:41Z","timestamp":1634467181000},"page":"293-303","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Gaze Based Implicit Intention Inference with Historical Information of Visual Attention for Human-Robot Interaction"],"prefix":"10.1007","author":[{"given":"Yujie","family":"Nie","sequence":"first","affiliation":[]},{"given":"Xin","family":"Ma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,10,18]]},"reference":[{"issue":"3","key":"27_CR1","doi-asserted-by":"publisher","first-page":"140","DOI":"10.1080\/10400435.2013.869703","volume":"26","author":"WYG Louie","year":"2014","unstructured":"Louie, W.Y.G., McColl, D., Nejat, G.: Acceptance and attitudes toward a human-like socially assistive robot by older adults. Assist. Technol. 26(3), 140\u2013150 (2014)","journal-title":"Assist. Technol."},{"key":"27_CR2","doi-asserted-by":"crossref","unstructured":"Trick, S., Koert, D., Peters, J.: Multimodal uncertainty reduction for intention recognition in human-robot interaction. arXiv preprint arXiv:1907.02426 (2019)","DOI":"10.1109\/IROS40897.2019.8968171"},{"key":"27_CR3","doi-asserted-by":"publisher","first-page":"103903","DOI":"10.1016\/j.engappai.2020.103903","volume":"95","author":"MC Bingol","year":"2020","unstructured":"Bingol, M.C., Aydogmus, O.: Performing predefined tasks using the human\u2013robot interaction on speech recognition for an industrial robot. Eng. Appl. Artif. Intell. 95, 103903 (2020)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"27_CR4","doi-asserted-by":"crossref","unstructured":"Pleshkova, S., Bekiarski, A.: Algorithm for motion management by interaction between mobile robot and human. In: 2019 II International Conference on High Technology for Sustainable Development (HiTech), pp. 1\u20134. IEEE (2019)","DOI":"10.1109\/HiTech48507.2019.9128233"},{"key":"27_CR5","doi-asserted-by":"crossref","unstructured":"Droeschel, D., St\u00fcckler, J., Behnke, S.: Learning to interpret pointing gestures with a time-of-flight camera. In: Proceedings of the 6th International Conference on Human-Robot Interaction, pp. 481\u2013488 (2011)","DOI":"10.1145\/1957656.1957822"},{"key":"27_CR6","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1016\/j.cviu.2016.03.004","volume":"149","author":"G Canal","year":"2016","unstructured":"Canal, G., Escalera, S., Angulo, C.: A real-time human-robot interaction system based on gestures for assistive scenarios. Comput. Vis. Image Underst. 149, 65\u201377 (2016)","journal-title":"Comput. Vis. Image Underst."},{"key":"27_CR7","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1016\/j.robot.2015.02.002","volume":"68","author":"W Sheng","year":"2015","unstructured":"Sheng, W., Du, J., Cheng, Q.: Robot semantic mapping through human activity recognition: a wearable sensing and computing approach. Robot. Auton. Syst. 68, 47\u201358 (2015)","journal-title":"Robot. Auton. Syst."},{"issue":"3","key":"27_CR8","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1109\/TSMCA.2010.2093883","volume":"41","author":"C Zhu","year":"2011","unstructured":"Zhu, C., Sheng, W.: Wearable sensor-based hand gesture and daily activity recognition for robot-assisted living. IEEE Trans. Syst. Man Cybern.-Part A: Syst. Hum. 41(3), 569\u2013573 (2011)","journal-title":"IEEE Trans. Syst. Man Cybern.-Part A: Syst. Hum."},{"key":"27_CR9","doi-asserted-by":"crossref","unstructured":"Rodomagoulakis, I., Kardaris, N., Pitsikalis, V.: Multimodal human action recognition in assistive human-robot interaction. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2702\u20132706. IEEE (2016)","DOI":"10.1109\/ICASSP.2016.7472168"},{"issue":"5","key":"27_CR10","doi-asserted-by":"publisher","first-page":"2524","DOI":"10.1109\/TFUZZ.2018.2809691","volume":"26","author":"L Chen","year":"2018","unstructured":"Chen, L., Zhou, M., Wu, M.: Three-layer weighted fuzzy support vector regression for emotional intention understanding in human\u2013robot interaction. IEEE Trans. Fuzzy Syst. 26(5), 2524\u20132538 (2018)","journal-title":"IEEE Trans. Fuzzy Syst."},{"key":"27_CR11","doi-asserted-by":"publisher","unstructured":"Yang, J.Y., Kwon, O.H., Lim, C.S., Kwon, D.S.: Human-robot interaction-based intention sharing of assistant robot for elderly people. In: Lee, S., Cho, H., Yoon, K.J., Lee, J. (eds.) Intelligent Autonomous Systems 12. AISC, vol. 194. pp. 283\u2013291. Springer, Heidelberg. (2013). https:\/\/doi.org\/10.1007\/978-3-642-33932-5_27","DOI":"10.1007\/978-3-642-33932-5_27"},{"key":"27_CR12","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1016\/j.neucom.2013.08.008","volume":"128","author":"YM Jang","year":"2014","unstructured":"Jang, Y.M., Mallipeddi, R., Lee, S., et al.: Human intention recognition based on eyeball movement pattern and pupil size variation. Neurocomputing 128, 421\u2013432 (2014)","journal-title":"Neurocomputing"},{"key":"27_CR13","doi-asserted-by":"publisher","first-page":"128","DOI":"10.1016\/j.cogsys.2018.12.010","volume":"54","author":"J Mi","year":"2019","unstructured":"Mi, J., Tang, S., Deng, Z.: Object affordance based multimodal fusion for natural human-robot interaction. Cogn. Syst. Res. 54, 128\u2013137 (2019)","journal-title":"Cogn. Syst. Res."},{"key":"27_CR14","doi-asserted-by":"crossref","unstructured":"Feng, Y., Chen, L., Wanjuan, S.U.: Gesture intention understanding based on depth and RGB data. In: 2018 37th Chinese Control Conference (CCC), pp. 9556\u20139559. IEEE (2018)","DOI":"10.23919\/ChiCC.2018.8483387"},{"key":"27_CR15","doi-asserted-by":"publisher","first-page":"105787","DOI":"10.1016\/j.asoc.2019.105787","volume":"85","author":"L Chen","year":"2019","unstructured":"Chen, L., Feng, Y., Maram, M.A.: Multi-SVM based Dempster-Shafer theory for gesture intention understanding using sparse coding feature. Appl. Soft Comput. 85, 105787 (2019)","journal-title":"Appl. Soft Comput."},{"key":"27_CR16","doi-asserted-by":"crossref","unstructured":"Park, S.M., Ko, K.E., Park, J.A.: study on hybrid model of HMMs and GMMs for mirror neuron system modeling using EEG signals. In: 2011 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE 2011), pp. 2752\u20132755. IEEE (2011)","DOI":"10.1109\/FUZZY.2011.6007503"},{"key":"27_CR17","doi-asserted-by":"publisher","first-page":"144","DOI":"10.1016\/j.patrec.2015.06.013","volume":"66","author":"JS Kang","year":"2015","unstructured":"Kang, J.S., Park, U., Gonuguntla, V.: Human implicit intent recognition based on the phase synchrony of EEG signals. Pattern Recogn. Lett. 66, 144\u2013152 (2015)","journal-title":"Pattern Recogn. Lett."},{"key":"27_CR18","doi-asserted-by":"crossref","unstructured":"Choi, J., Lee, S.J., Kim, S.J.: Detecting voluntary gait initiation\/termination intention using EEG. In: 2018 6th International Conference on Brain-Computer Interface (BCI), pp. 1\u20133. IEEE (2018)","DOI":"10.1109\/IWW-BCI.2018.8311532"},{"issue":"3","key":"27_CR19","first-page":"480","volume":"33","author":"MR Ahsan","year":"2009","unstructured":"Ahsan, M.R., Ibrahimy, M.I., Khalifa, O.O.: EMG signal classification for human computer interaction: a review. Eur. J. Sci. Res. 33(3), 480\u2013501 (2009)","journal-title":"Eur. J. Sci. Res."},{"key":"27_CR20","doi-asserted-by":"crossref","unstructured":"Li, T., Yang, J., Bai, D.: A new directional intention identification approach for intelligent wheelchair based on fusion of EOG signal and eye movement signal. In: 2018 IEEE International Conference on Intelligence and Safety for Robotics (ISR), pp. 470\u2013474. IEEE (2018)","DOI":"10.1109\/IISR.2018.8535709"},{"key":"27_CR21","doi-asserted-by":"crossref","unstructured":"Sy, A.C., Bugtai, N.T.: Velocity and acceleration induced response to bicep EMG signal threshold for motion intention detection. In: 2014 International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment and Management (HNICEM), pp. 1\u20136. IEEE (2014)","DOI":"10.1109\/HNICEM.2014.7016198"},{"key":"27_CR22","doi-asserted-by":"crossref","unstructured":"Fernandes, P.N., Figueredo, J., Moreira, L.: EMG-based motion intention recognition for controlling a powered knee orthosis. In: 2019 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC), pp. 1\u20136. IEEE (2019)","DOI":"10.1109\/ICARSC.2019.8733628"},{"issue":"3","key":"27_CR23","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1007\/s00221-006-0804-0","volume":"179","author":"N Mennie","year":"2007","unstructured":"Mennie, N., Hayhoe, M., Sullivan, B.: Look-ahead fixations: anticipatory eye movements in natural tasks. Exp. Brain Res. 179(3), 427\u2013442 (2007)","journal-title":"Exp. Brain Res."},{"key":"27_CR24","unstructured":"Admoni, H., Srinivasa, S.: Predicting user intent through eye gaze for shared autonomy. In: 2016 AAAI Fall Symposium Series (2016)"},{"issue":"11","key":"27_CR25","doi-asserted-by":"publisher","first-page":"1311","DOI":"10.1068\/p2935","volume":"28","author":"M Land","year":"1999","unstructured":"Land, M., Mennie, N., Rusted, J.: The roles of vision and eye movements in the control of activities of daily living. Perception 28(11), 1311\u20131328 (1999)","journal-title":"Perception"},{"issue":"4","key":"27_CR26","doi-asserted-by":"publisher","first-page":"437","DOI":"10.1109\/THMS.2017.2647882","volume":"47","author":"S Li","year":"2017","unstructured":"Li, S., Zhang, X.: Implicit intention communication in human\u2013robot interaction through visual behavior studies. IEEE Trans. Hum.-Mach. Syst. 47(4), 437\u2013448 (2017)","journal-title":"IEEE Trans. Hum.-Mach. Syst."},{"issue":"2","key":"27_CR27","doi-asserted-by":"publisher","first-page":"B25","DOI":"10.1016\/S0010-0277(98)00009-2","volume":"66","author":"AS Meyer","year":"1998","unstructured":"Meyer, A.S., Sleiderink, A.M., Levelt, W.J.M.: Viewing and naming objects: eye movements during noun phrase production. Cognition 66(2), B25\u2013B33 (1998)","journal-title":"Cognition"},{"issue":"25\u201326","key":"27_CR28","doi-asserted-by":"publisher","first-page":"3587","DOI":"10.1016\/S0042-6989(01)00245-0","volume":"41","author":"JB Pelz","year":"2001","unstructured":"Pelz, J.B., Canosa, R.: Oculomotor behavior and perceptual strategies in complex tasks. Vision. Res. 41(25\u201326), 3587\u20133596 (2001)","journal-title":"Vision. Res."},{"key":"27_CR29","first-page":"1049","volume":"6","author":"CM Huang","year":"2015","unstructured":"Huang, C.M., Andrist, S., Saupp\u00e9, A.: Using gaze patterns to predict task intent in collaboration. Front. Psychol. 6, 1049 (2015)","journal-title":"Front. Psychol."},{"issue":"2","key":"27_CR30","first-page":"185","volume":"2","author":"PS Gajwani","year":"2010","unstructured":"Gajwani, P.S., Chhabria, S.A.: Eye motion tracking for wheelchair control. Int. J. Inf. Technol. 2(2), 185\u2013187 (2010)","journal-title":"Int. J. Inf. Technol."},{"key":"27_CR31","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1007\/978-1-4471-4784-8_6","volume-title":"Eye Gaze in Intelligent User Interfaces","author":"R Ishii","year":"2013","unstructured":"Ishii, R., Ooko, R., Nakano, Y.I.: Effectiveness of gaze-based engagement estimation in conversational agents. In: Nakano, Y., Conati, C., Bader, T. (eds.) Eye Gaze in Intelligent User Interfaces, pp. 85\u2013110. Springer, London (2013). https:\/\/doi.org\/10.1007\/978-1-4471-4784-8_6"},{"issue":"3","key":"27_CR32","first-page":"273","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes, C., Vapnik, V.: Support vector machine. Mach. Learn. 20(3), 273\u2013297 (1995)","journal-title":"Mach. Learn."}],"container-title":["Lecture Notes in Computer Science","Intelligent Robotics and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-89134-3_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,10,17]],"date-time":"2021-10-17T11:01:03Z","timestamp":1634468463000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-89134-3_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030891336","9783030891343"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-89134-3_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"18 October 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIRA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Robotics and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Yantai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 October 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 October 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icira2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.icira2021.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}