{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T12:22:07Z","timestamp":1764937327814,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,10,18]],"date-time":"2021-10-18T00:00:00Z","timestamp":1634515200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,10,18]]},"DOI":"10.1145\/3462244.3479937","type":"proceedings-article","created":{"date-parts":[[2021,10,15]],"date-time":"2021-10-15T14:41:47Z","timestamp":1634308907000},"page":"461-469","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Inclusive Voice Interaction Techniques for Creative Object Positioning"],"prefix":"10.1145","author":[{"given":"Farkhandah","family":"Aziz","sequence":"first","affiliation":[{"name":"Birmingham City University, United Kingdom"}]},{"given":"Chris","family":"Creed","sequence":"additional","affiliation":[{"name":"Birmingham City University, United Kingdom"}]},{"given":"Maite","family":"Frutos-Pascual","sequence":"additional","affiliation":[{"name":"Birmingham City University, United Kingdom"}]},{"given":"Ian","family":"Williams","sequence":"additional","affiliation":[{"name":"Birmingham City University, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2021,10,18]]},"reference":[{"volume-title":"Retrieved","year":"1999","key":"e_1_3_2_2_1_1","unstructured":"Adobe. 1999 . Photoshop apps - desktop, mobile, and tablet | Photoshop.com . Retrieved May 31, 2021 from https:\/\/www.photoshop.com\/en Adobe. 1999. Photoshop apps - desktop, mobile, and tablet | Photoshop.com. Retrieved May 31, 2021 from https:\/\/www.photoshop.com\/en"},{"key":"e_1_3_2_2_2_1","volume-title":"Retrieved","author":"Adobe Inc.","year":"2012","unstructured":"Adobe Inc. 2012 . Adobe Illustrator CS6: Industry-leading vector graphics software . Retrieved May 31, 2021 from https:\/\/www.adobe.com\/uk\/products\/illustrator.html Adobe Inc. 2012. Adobe Illustrator CS6: Industry-leading vector graphics software. Retrieved May 31, 2021 from https:\/\/www.adobe.com\/uk\/products\/illustrator.html"},{"key":"e_1_3_2_2_3_1","volume-title":"Retrieved","author":"Adobe Inc.","year":"2021","unstructured":"Adobe Inc. 2021 . Adobe XD | Fast & Powerful UI\/UX Design & Collaboration Tool . Retrieved May 31, 2021 from https:\/\/www.adobe.com\/uk\/products\/xd.html Adobe Inc. 2021. Adobe XD | Fast & Powerful UI\/UX Design & Collaboration Tool. Retrieved May 31, 2021 from https:\/\/www.adobe.com\/uk\/products\/xd.html"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.14236\/ewic\/HCI2007.42"},{"key":"e_1_3_2_2_5_1","first-page":"114","article-title":"Determining what individual SUS scores mean: adding an adjective rating scale","author":"Bangor Aaron","year":"2009","unstructured":"Aaron Bangor , Philip Kortum , and James Miller . 2009 . Determining what individual SUS scores mean: adding an adjective rating scale . Journal of usability studies. 114 \u2013 123 . Aaron Bangor, Philip Kortum, and James Miller. 2009. Determining what individual SUS scores mean: adding an adjective rating scale. Journal of usability studies. 114\u2013123.","journal-title":"Journal of usability studies."},{"volume-title":"The role of voice input for human-machine communication","author":"Cohen Philip R","key":"e_1_3_2_2_6_1","unstructured":"Philip R Cohen and Sharon L Oviatt . 1995. The role of voice input for human-machine communication . Proceedings of National Academy of Sciences . 9921\u20139927. DOI:https:\/\/doi.org\/10.1073\/pnas.92.22.9921 Philip R Cohen and Sharon L Oviatt. 1995. The role of voice input for human-machine communication. Proceedings of National Academy of Sciences. 9921\u20139927. DOI:https:\/\/doi.org\/10.1073\/pnas.92.22.9921"},{"volume-title":"Assistive technology for disabled visual artists: exploring the impact of digital technologies on artistic practice","author":"Creed Chris","key":"e_1_3_2_2_7_1","unstructured":"Chris Creed . 2018. Assistive technology for disabled visual artists: exploring the impact of digital technologies on artistic practice . Disability and Society . 1103\u20131119. DOI:https:\/\/doi.org\/10.1080\/09687599.2018.1469400 Chris Creed. 2018. Assistive technology for disabled visual artists: exploring the impact of digital technologies on artistic practice. Disability and Society. 1103\u20131119. DOI:https:\/\/doi.org\/10.1080\/09687599.2018.1469400"},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661334.2661386"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376196"},{"key":"e_1_3_2_2_10_1","volume-title":"Proceedings of IHCI 2011-IADIS conference on Interfaces and Human Computer Interaction. Retrieved","author":"Duarte Carlos","year":"2011","unstructured":"Carlos Duarte and Joana Neca . 2011 . Evaluation of Gestural Interaction with and without Voice Commands . In Proceedings of IHCI 2011-IADIS conference on Interfaces and Human Computer Interaction. Retrieved January 27, 2021 from https:\/\/www.researchgate.net\/publication\/256007049 Carlos Duarte and Joana Neca. 2011. Evaluation of Gestural Interaction with and without Voice Commands. In Proceedings of IHCI 2011-IADIS conference on Interfaces and Human Computer Interaction. Retrieved January 27, 2021 from https:\/\/www.researchgate.net\/publication\/256007049"},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.4135\/9781412983419"},{"key":"e_1_3_2_2_12_1","unstructured":"Arno Gourdol Laurence Nigay Daniel Salber and Joelle Coutaz. 1992. Two Case Studies of Software Architecture for Multimodal Interactive Systems: VoicePaint and a Voice-enabled Graphical Notebook. Engineering for Human-Computer Interaction. 271\u201384.  Arno Gourdol Laurence Nigay Daniel Salber and Joelle Coutaz. 1992. Two Case Studies of Software Architecture for Multimodal Interactive Systems: VoicePaint and a Voice-enabled Graphical Notebook. Engineering for Human-Computer Interaction. 271\u201384."},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/1322192.1322225"},{"volume-title":"ASSETS\u201907: Proceedings of the Ninth International ACM SIGACCESS Conference on Computers and Accessibility. 27\u201334","author":"Harada Susumu","key":"e_1_3_2_2_14_1","unstructured":"Susumu Harada , Jacob O. Wobbrock , and James A. Landay . 2007. VoiceDraw: A hands-free voice-driven drawing application for people with motor impairments . In ASSETS\u201907: Proceedings of the Ninth International ACM SIGACCESS Conference on Computers and Accessibility. 27\u201334 . DOI:https:\/\/doi.org\/10.1145\/1296843.1296850 Susumu Harada, Jacob O. Wobbrock, and James A. Landay. 2007. VoiceDraw: A hands-free voice-driven drawing application for people with motor impairments. In ASSETS\u201907: Proceedings of the Ninth International ACM SIGACCESS Conference on Computers and Accessibility. 27\u201334. DOI:https:\/\/doi.org\/10.1145\/1296843.1296850"},{"key":"e_1_3_2_2_15_1","volume-title":"Proceedings of the SIGCHI conference on Human Factors in Computing Systems. 347\u2013356","author":"Harada Susumu","year":"1870","unstructured":"Susumu Harada , Jacob O. Wobbrock , Jonathan Malkin , Jeff A. Bilmes , and James A. Landay . 2009. Longitudinal study of people learning to use continuous voice-based cursor control . In Proceedings of the SIGCHI conference on Human Factors in Computing Systems. 347\u2013356 . DOI:https:\/\/doi.org\/10.1145\/15 1870 1.1518757 Susumu Harada, Jacob O. Wobbrock, Jonathan Malkin, Jeff A. Bilmes, and James A. Landay. 2009. Longitudinal study of people learning to use continuous voice-based cursor control. In Proceedings of the SIGCHI conference on Human Factors in Computing Systems. 347\u2013356. DOI:https:\/\/doi.org\/10.1145\/1518701.1518757"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/67449.67496"},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.3115\/991250.991262"},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-21657-2_7"},{"key":"e_1_3_2_2_19_1","volume-title":"Karn","author":"Jacob Robert J.K.","year":"2003","unstructured":"Robert J.K. Jacob and Keith S . Karn . 2003 . Eye Tracking in human-computer interaction and usability research: Ready to deliver the promises. In The Mind's eye. North-Holland . 573\u2013605. DOI:https:\/\/doi.org\/10.1016\/B978-044451020-4\/50031-1 Robert J.K. Jacob and Keith S. Karn. 2003. Eye Tracking in human-computer interaction and usability research: Ready to deliver the promises. In The Mind's eye. North-Holland. 573\u2013605. DOI:https:\/\/doi.org\/10.1016\/B978-044451020-4\/50031-1"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/354324.354334"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/638249.638258"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/1983302.1983311"},{"key":"e_1_3_2_2_23_1","volume-title":"Karimullah","author":"Sears Andrew","year":"2002","unstructured":"Andrew Sears , Min Lin , and Azfar S . Karimullah . 2002 . Speech-based cursor control: understanding the effects of target size, cursor speed, and command selection. Universal Access in the Information Society . 30\u201343. DOI:https:\/\/doi.org\/10.1007\/s10209-002-0034-6 Andrew Sears, Min Lin, and Azfar S. Karimullah. 2002. Speech-based cursor control: understanding the effects of target size, cursor speed, and command selection. Universal Access in the Information Society. 30\u201343. DOI:https:\/\/doi.org\/10.1007\/s10209-002-0034-6"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/638249.638282"},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300562"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/2470654.2481301"},{"volume-title":"Team Chat & Collaboration. Microsoft Teams. Retrieved","year":"2021","key":"e_1_3_2_2_27_1","unstructured":"Microsoft. 2021 . Microsoft Teams | Group Chat , Team Chat & Collaboration. Microsoft Teams. Retrieved May 31, 2021 from https:\/\/www.microsoft.com\/en-gb\/microsoft-teams\/group-chat-software Microsoft. 2021. Microsoft Teams | Group Chat, Team Chat & Collaboration. Microsoft Teams. Retrieved May 31, 2021 from https:\/\/www.microsoft.com\/en-gb\/microsoft-teams\/group-chat-software"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/ROMAN.1995.531944"},{"key":"e_1_3_2_2_29_1","article-title":"An empirical study: Adding voice input to a graphical editor","author":"Pausch Randy","year":"1991","unstructured":"Randy Pausch and James H. Leatherby . 1991 . An empirical study: Adding voice input to a graphical editor . In Journal of the American Voice Input\/Output Society. Retrieved January 27, 2021 from http:\/\/citeseerx.ist.psu.edu\/viewdoc\/summary?doi=10.1.1.49.4829 Randy Pausch and James H. Leatherby. 1991. An empirical study: Adding voice input to a graphical editor. In Journal of the American Voice Input\/Output Society. Retrieved January 27, 2021 from http:\/\/citeseerx.ist.psu.edu\/viewdoc\/summary?doi=10.1.1.49.4829","journal-title":"Journal of the American Voice Input\/Output Society. Retrieved"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/317561.317571"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"crossref","unstructured":"Samuel S Shapiro and Martin B Wilk. 1965. An Analysis of Variance Test for Normality (Complete Samples). Biometrika. 591\u2013611. DOI:https:\/\/doi.org\/10.2307\/2333709  Samuel S Shapiro and Martin B Wilk. 1965. An Analysis of Variance Test for Normality (Complete Samples). Biometrika. 591\u2013611. DOI:https:\/\/doi.org\/10.2307\/2333709","DOI":"10.1093\/biomet\/52.3-4.591"},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3301275.3302292"},{"volume-title":"On the Euclidean distance of images","author":"Wang Liwei","key":"e_1_3_2_2_33_1","unstructured":"Liwei Wang , Yan Zhang , and Jufu Feng . 2005. On the Euclidean distance of images . IEEE transactions on pattern analysis and machine intelligence. 1334\u20131339. DOI:https:\/\/doi.org\/10.1109\/TPAMI.2005.165 Liwei Wang, Yan Zhang, and Jufu Feng. 2005. On the Euclidean distance of images. IEEE transactions on pattern analysis and machine intelligence. 1334\u20131339. DOI:https:\/\/doi.org\/10.1109\/TPAMI.2005.165"},{"key":"e_1_3_2_2_34_1","volume-title":"Fisher's Least Signiflcant Difierence (LSD) Test. Encyclopedia of research design. 840\u2013853. Retrieved","author":"Williams Lynne J","year":"2021","unstructured":"Lynne J Williams and Herve Abdi . 2010. Fisher's Least Signiflcant Difierence (LSD) Test. Encyclopedia of research design. 840\u2013853. Retrieved January 27, 2021 from https:\/\/www.researchgate.net\/publication\/242181775 Lynne J Williams and Herve Abdi. 2010. Fisher's Least Signiflcant Difierence (LSD) Test. Encyclopedia of research design. 840\u2013853. Retrieved January 27, 2021 from https:\/\/www.researchgate.net\/publication\/242181775"},{"key":"e_1_3_2_2_35_1","volume-title":"Speech-Based Navigation: Improving Grid-Based Solutions. In IFIP conference on Human-Computer Interaction. 50\u201362","author":"Zhu Shaojian","year":"2009","unstructured":"Shaojian Zhu , Yao Ma , Jinjuan Feng , and Andrew Sears . 2009 . Speech-Based Navigation: Improving Grid-Based Solutions. In IFIP conference on Human-Computer Interaction. 50\u201362 . Springer, Berlin, Heidelberg Shaojian Zhu, Yao Ma, Jinjuan Feng, and Andrew Sears. 2009. Speech-Based Navigation: Improving Grid-Based Solutions. In IFIP conference on Human-Computer Interaction. 50\u201362. Springer, Berlin, Heidelberg"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"crossref","unstructured":"Shaojian Zhu Andrew Sears and Jinjuan Feng. 2010. Investigating grid-based navigation:The impact of physical disability. ACM Transactions on Accessible Computing. 1\u201330. DOI:https:\/\/doi.org\/10.1145\/1838562.1838565  Shaojian Zhu Andrew Sears and Jinjuan Feng. 2010. Investigating grid-based navigation:The impact of physical disability. ACM Transactions on Accessible Computing. 1\u201330. DOI:https:\/\/doi.org\/10.1145\/1838562.1838565","DOI":"10.1145\/1838562.1838565"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1080\/00220973.1993.9943832"},{"volume-title":"Retrieved","year":"2019","key":"e_1_3_2_2_38_1","unstructured":"Zoom. 2019 . Video Conferencing, Web Conferencing, Webinars, Screen Sharing - Zoom. Zoom (2019) . Retrieved January 25, 2021 from https:\/\/www.zoom.us\/ Zoom. 2019. Video Conferencing, Web Conferencing, Webinars, Screen Sharing - Zoom. Zoom (2019). Retrieved January 25, 2021 from https:\/\/www.zoom.us\/"},{"key":"e_1_3_2_2_39_1","unstructured":"Central London RSI Support Group - Home | Facebook. Retrieved January 25 2021 from https:\/\/www.facebook.com\/CentralLondonRsiSupportGroup\/  Central London RSI Support Group - Home | Facebook. Retrieved January 25 2021 from https:\/\/www.facebook.com\/CentralLondonRsiSupportGroup\/"},{"key":"e_1_3_2_2_40_1","volume-title":"Retrieved","author":"Loud Designing Out","year":"2021","unstructured":"Designing Out Loud : Announcing Support for macOS Voice Control in Adobe XD. 2020 . Retrieved January 25, 2021 from https:\/\/blog.adobe.com\/en\/2020\/02\/11\/announcing-mac-os-voice-control-adobe-xd.html?scid=fac788f5-fe6f-4be4-a960-871ac58b5f30&mv=social&mv2=owned_social Designing Out Loud: Announcing Support for macOS Voice Control in Adobe XD. 2020. Retrieved January 25, 2021 from https:\/\/blog.adobe.com\/en\/2020\/02\/11\/announcing-mac-os-voice-control-adobe-xd.html?scid=fac788f5-fe6f-4be4-a960-871ac58b5f30&mv=social&mv2=owned_social"},{"key":"e_1_3_2_2_41_1","unstructured":"Using the Web Speech API - Web APIs | MDN. Retrieved May 31 2021 from https:\/\/developer.mozilla.org\/en-US\/docs\/Web\/API\/Web_Speech_API\/Using_the_Web_Speech_API  Using the Web Speech API - Web APIs | MDN. Retrieved May 31 2021 from https:\/\/developer.mozilla.org\/en-US\/docs\/Web\/API\/Web_Speech_API\/Using_the_Web_Speech_API"},{"key":"e_1_3_2_2_42_1","unstructured":"Facebook - Log In or Sign Up. Retrieved May 31 2021 from https:\/\/www.facebook.com\/  Facebook - Log In or Sign Up. Retrieved May 31 2021 from https:\/\/www.facebook.com\/"},{"volume-title":"Retrieved","year":"2021","key":"e_1_3_2_2_43_1","unstructured":"LinkedIn : Log In or Sign Up . Retrieved May 31, 2021 from https:\/\/www.linkedin.com\/ LinkedIn: Log In or Sign Up. Retrieved May 31, 2021 from https:\/\/www.linkedin.com\/"},{"key":"e_1_3_2_2_44_1","unstructured":"Instagram. Retrieved May 31 2021 from https:\/\/www.instagram.com\/  Instagram. Retrieved May 31 2021 from https:\/\/www.instagram.com\/"}],"event":{"name":"ICMI '21: INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"Montr\u00e9al QC Canada","acronym":"ICMI '21"},"container-title":["Proceedings of the 2021 International Conference on Multimodal Interaction"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3462244.3479937","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3462244.3479937","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:48:55Z","timestamp":1750193335000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3462244.3479937"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,18]]},"references-count":44,"alternative-id":["10.1145\/3462244.3479937","10.1145\/3462244"],"URL":"https:\/\/doi.org\/10.1145\/3462244.3479937","relation":{},"subject":[],"published":{"date-parts":[[2021,10,18]]},"assertion":[{"value":"2021-10-18","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}