{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T13:27:32Z","timestamp":1773840452640,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T00:00:00Z","timestamp":1651190400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,4,29]]},"DOI":"10.1145\/3491102.3502134","type":"proceedings-article","created":{"date-parts":[[2022,4,28]],"date-time":"2022-04-28T16:34:58Z","timestamp":1651163698000},"page":"1-14","source":"Crossref","is-referenced-by-count":12,"title":["Integrating Gaze and Speech for Enabling Implicit Interactions"],"prefix":"10.1145","author":[{"given":"Anam Ahmad","family":"Khan","sequence":"first","affiliation":[{"name":"The University of Melbourne, Australia"}]},{"given":"Joshua","family":"Newn","sequence":"additional","affiliation":[{"name":"Computing and Information Systems, The University of Melbourne, Australia"}]},{"given":"James","family":"Bailey","sequence":"additional","affiliation":[{"name":"Department of Computing and Information Systems, The University of Melbourne, Australia"}]},{"given":"Eduardo","family":"Velloso","sequence":"additional","affiliation":[{"name":"School of Computing and Information Systems, University of Melbourne, Australia"}]}],"member":"320","published-online":{"date-parts":[[2022,4,29]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"2021. Hypothesis. https:\/\/web.hypothes.is\/ 2021. Hypothesis. https:\/\/web.hypothes.is\/"},{"key":"e_1_3_2_2_2_1","unstructured":"2021. reflect-in-seesaw. https:\/\/chrome.google.com\/webstore\/detail\/reflect-in-seesaw-extensi\/lhgiigkiddoalobhmmcpdhddlccindjj 2021. reflect-in-seesaw. https:\/\/chrome.google.com\/webstore\/detail\/reflect-in-seesaw-extensi\/lhgiigkiddoalobhmmcpdhddlccindjj"},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2016.2634527"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/1647314.1647350"},{"key":"e_1_3_2_2_5_1","volume-title":"Chart-text: A fully automated chart image descriptor. arXiv preprint arXiv:1812.10636(2018).","author":"Balaji Abhijit","year":"2018","unstructured":"Abhijit Balaji , Thuvaarakkesh Ramanathan , and Venkateshwarlu Sonathi . 2018 . Chart-text: A fully automated chart image descriptor. arXiv preprint arXiv:1812.10636(2018). Abhijit Balaji, Thuvaarakkesh Ramanathan, and Venkateshwarlu Sonathi. 2018. Chart-text: A fully automated chart image descriptor. arXiv preprint arXiv:1812.10636(2018)."},{"key":"e_1_3_2_2_6_1","volume-title":"The usability of speech and eye gaze as a multimodal interface for a word processor. Speech Technologies","author":"Beelders TR","year":"2011","unstructured":"TR Beelders and PJ Blignaut . 2011. The usability of speech and eye gaze as a multimodal interface for a word processor. Speech Technologies ( 2011 ), 386\u2013404. TR Beelders and PJ Blignaut. 2011. The usability of speech and eye gaze as a multimodal interface for a word processor. Speech Technologies (2011), 386\u2013404."},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/800250.807503"},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379157.3388929"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1223"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2449396.2449406"},{"key":"e_1_3_2_2_11_1","unstructured":"Andrew\u00a0M Dai Christopher Olah and Quoc\u00a0V Le. 2015. Document embedding with paragraph vectors. arXiv preprint arXiv:1507.07998(2015). Andrew\u00a0M Dai Christopher Olah and Quoc\u00a0V Le. 2015. Document embedding with paragraph vectors. arXiv preprint arXiv:1507.07998(2015)."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1002\/meet.14504201151"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-04962-0_53"},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2014.08.003"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/2663204.2663277"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/2499474.2499481"},{"key":"e_1_3_2_2_17_1","volume-title":"Proceedings of SigDial 2013(2013)","author":"Kennington Casey","year":"2013","unstructured":"Casey Kennington , Spyridon Kousidis , and David Schlangen . 2013 . Interpreting situated dialogue utterances: an update model that uses speech, gaze, and gesture information . Proceedings of SigDial 2013(2013) . Casey Kennington, Spyridon Kousidis, and David Schlangen. 2013. Interpreting situated dialogue utterances: an update model that uses speech, gaze, and gesture information. Proceedings of SigDial 2013(2013)."},{"key":"e_1_3_2_2_18_1","volume-title":"Using voice note-taking to promote learners","author":"Khan Anam\u00a0Ahmad","year":"2012","unstructured":"Anam\u00a0Ahmad Khan , Sadia Nawaz , Joshua Newn , Jason\u00a0 M. Lodge , James Bailey , and Eduardo Velloso . 2020. Using voice note-taking to promote learners \u2019 conceptual understanding. arxiv: 2012 .02927 Anam\u00a0Ahmad Khan, Sadia Nawaz, Joshua Newn, Jason\u00a0M. Lodge, James Bailey, and Eduardo Velloso. 2020. Using voice note-taking to promote learners\u2019 conceptual understanding. arxiv:2012.02927"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3453988"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300562"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3279972.3279976"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/CITSM.2016.7577578"},{"key":"e_1_3_2_2_23_1","volume-title":"International conference on machine learning. PMLR, 1188\u20131196","author":"Le Quoc","year":"2014","unstructured":"Quoc Le and Tomas Mikolov . 2014 . Distributed representations of sentences and documents . In International conference on machine learning. PMLR, 1188\u20131196 . Quoc Le and Tomas Mikolov. 2014. Distributed representations of sentences and documents. In International conference on machine learning. PMLR, 1188\u20131196."},{"key":"e_1_3_2_2_24_1","volume-title":"Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1\u201315","author":"Jia-Jun Li Toby","year":"2021","unstructured":"Toby Jia-Jun Li , Lindsay Popowski , Tom Mitchell , and Brad\u00a0 A Myers . 2021 . Screen2Vec: Semantic Embedding of GUI Screens and GUI Components . In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1\u201315 . Toby Jia-Jun Li, Lindsay Popowski, Tom Mitchell, and Brad\u00a0A Myers. 2021. Screen2Vec: Semantic Embedding of GUI Screens and GUI Components. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1\u201315."},{"key":"e_1_3_2_2_25_1","unstructured":"Scott\u00a0M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Advances in neural information processing systems. 4765\u20134774. Scott\u00a0M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Advances in neural information processing systems. 4765\u20134774."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376479"},{"key":"e_1_3_2_2_28_1","volume-title":"Patterns for How Users Overcome Obstacles in Voice User Interfaces","author":"Myers Chelsea","unstructured":"Chelsea Myers , Anushay Furqan , Jessica Nebolsky , Karina Caro , and Jichen Zhu . 2018. Patterns for How Users Overcome Obstacles in Voice User Interfaces . Association for Computing Machinery , New York, NY, USA , 1\u20137. https:\/\/doi.org\/10.1145\/3173574.3173580 Chelsea Myers, Anushay Furqan, Jessica Nebolsky, Karina Caro, and Jichen Zhu. 2018. Patterns for How Users Overcome Obstacles in Voice User Interfaces. Association for Computing Machinery, New York, NY, USA, 1\u20137. https:\/\/doi.org\/10.1145\/3173574.3173580"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2858036.2858137"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"crossref","unstructured":"Panupong Pasupat Tian-Shun Jiang Evan\u00a0Zheran Liu Kelvin Guu and Percy Liang. 2018. Mapping natural language commands to web elements. arXiv preprint arXiv:1808.09132(2018). Panupong Pasupat Tian-Shun Jiang Evan\u00a0Zheran Liu Kelvin Guu and Percy Liang. 2018. Mapping natural language commands to web elements. arXiv preprint arXiv:1808.09132(2018).","DOI":"10.18653\/v1\/D18-1540"},{"key":"e_1_3_2_2_31_1","volume-title":"Cognitive effort during note taking. Applied cognitive psychology 19, 3","author":"Piolat Annie","year":"2005","unstructured":"Annie Piolat , Thierry Olive , and Ronald\u00a0 T Kellogg . 2005. Cognitive effort during note taking. Applied cognitive psychology 19, 3 ( 2005 ), 291\u2013312. Annie Piolat, Thierry Olive, and Ronald\u00a0T Kellogg. 2005. Cognitive effort during note taking. Applied cognitive psychology 19, 3 (2005), 291\u2013312."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174214"},{"key":"e_1_3_2_2_33_1","volume-title":"Proceedings of the 13th international conference on Intelligent user interfaces. 20\u201329","author":"Prasov Zahar","year":"2008","unstructured":"Zahar Prasov and Joyce\u00a0 Y Chai . 2008 . What\u2019s in a gaze? The role of eye-gaze in reference resolution in multimodal conversational interfaces . In Proceedings of the 13th international conference on Intelligent user interfaces. 20\u201329 . Zahar Prasov and Joyce\u00a0Y Chai. 2008. What\u2019s in a gaze? The role of eye-gaze in reference resolution in multimodal conversational interfaces. In Proceedings of the 13th international conference on Intelligent user interfaces. 20\u201329."},{"key":"e_1_3_2_2_34_1","volume-title":"Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084(2019).","author":"Reimers Nils","year":"2019","unstructured":"Nils Reimers and Iryna Gurevych . 2019 . Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084(2019). Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084(2019)."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/355017.355028"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3204493.3208338"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"crossref","first-page":"595","DOI":"10.1093\/jcr\/ucy011","article-title":"Word of mouth versus word of mouse: Speaking about a brand connects you to it more than writing does","volume":"45","author":"Shen Hao","year":"2018","unstructured":"Hao Shen and Jaideep Sengupta . 2018 . Word of mouth versus word of mouse: Speaking about a brand connects you to it more than writing does . Journal of Consumer Research 45 , 3 (2018), 595 \u2013 614 . https:\/\/doi.org\/10.1093\/jcr\/ucy011 Hao Shen and Jaideep Sengupta. 2018. Word of mouth versus word of mouse: Speaking about a brand connects you to it more than writing does. Journal of Consumer Research 45, 3 (2018), 595\u2013614. https:\/\/doi.org\/10.1093\/jcr\/ucy011","journal-title":"Journal of Consumer Research"},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/2663204.2663251"},{"key":"e_1_3_2_2_39_1","volume-title":"User interaction in hands-free gaming: A comparative study of gaze-voice and touchscreen interface control. Turkish Journal of Electrical Engineering and Computer Sciences 26 (07","author":"Uludagli Cagkan","year":"2018","unstructured":"Cagkan Uludagli and Cengiz Acarturk . 2018. User interaction in hands-free gaming: A comparative study of gaze-voice and touchscreen interface control. Turkish Journal of Electrical Engineering and Computer Sciences 26 (07 2018 ). https:\/\/doi.org\/10.3906\/elk-1710-128 Cagkan Uludagli and Cengiz Acarturk. 2018. User interaction in hands-free gaming: A comparative study of gaze-voice and touchscreen interface control. Turkish Journal of Electrical Engineering and Computer Sciences 26 (07 2018). https:\/\/doi.org\/10.3906\/elk-1710-128"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/1983302.1983311"},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/1983302.1983311"},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2700648.2811369"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298935"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"crossref","unstructured":"Bryan Wang Gang Li Xin Zhou Zhourong Chen Tovi Grossman and Yang Li. 2021. Screen2Words: Automatic Mobile UI Summarization with Multimodal Learning. arXiv preprint arXiv:2108.03353(2021). Bryan Wang Gang Li Xin Zhou Zhourong Chen Tovi Grossman and Yang Li. 2021. Screen2Words: Automatic Mobile UI Summarization with Multimodal Learning. arXiv preprint arXiv:2108.03353(2021).","DOI":"10.1145\/3472749.3474765"},{"key":"e_1_3_2_2_45_1","volume-title":"Recent trends in deep learning based natural language processing. ieee Computational intelligenCe magazine 13, 3","author":"Young Tom","year":"2018","unstructured":"Tom Young , Devamanyu Hazarika , Soujanya Poria , and Erik Cambria . 2018. Recent trends in deep learning based natural language processing. ieee Computational intelligenCe magazine 13, 3 ( 2018 ), 55\u201375. Tom Young, Devamanyu Hazarika, Soujanya Poria, and Erik Cambria. 2018. Recent trends in deep learning based natural language processing. ieee Computational intelligenCe magazine 13, 3 (2018), 55\u201375."}],"event":{"name":"CHI '22: CHI Conference on Human Factors in Computing Systems","location":"New Orleans LA USA","acronym":"CHI '22","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["CHI Conference on Human Factors in Computing Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3491102.3502134","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3491102.3502134","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:23Z","timestamp":1750188683000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3491102.3502134"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,29]]},"references-count":44,"alternative-id":["10.1145\/3491102.3502134","10.1145\/3491102"],"URL":"https:\/\/doi.org\/10.1145\/3491102.3502134","relation":{},"subject":[],"published":{"date-parts":[[2022,4,29]]}}}