{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T22:22:22Z","timestamp":1776118942663,"version":"3.50.1"},"reference-count":70,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2020,12,17]],"date-time":"2020-12-17T00:00:00Z","timestamp":1608163200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["1839379"],"award-info":[{"award-number":["1839379"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"national science foundation","award":["1822819"],"award-info":[{"award-number":["1822819"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2020,12,17]]},"abstract":"<jats:p>This paper presents a holistic system to scale up the teaching and learning of vocabulary words of American Sign Language (ASL). The system leverages the most recent mixed-reality technology to allow the user to perceive her own hands in an immersive learning environment with first- and third-person views for motion demonstration and practice. Precise motion sensing is used to record and evaluate motion, providing real-time feedback tailored to the specific learner. As part of this evaluation, learner motions are matched to features derived from the Hamburg Notation System (HNS) developed by sign-language linguists. We develop a prototype to evaluate the efficacy of mixed-reality-based interactive motion teaching. Results with 60 participants show a statistically significant improvement in learning ASL signs when using our system, in comparison to traditional desktop-based, non-interactive learning. We expect this approach to ultimately allow teaching and guided practice of thousands of signs.<\/jats:p>","DOI":"10.1145\/3432211","type":"journal-article","created":{"date-parts":[[2020,12,18]],"date-time":"2020-12-18T15:39:14Z","timestamp":1608305954000},"page":"1-27","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":41,"title":["Teaching American Sign Language in Mixed Reality"],"prefix":"10.1145","volume":"4","author":[{"given":"Qijia","family":"Shao","sequence":"first","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]},{"given":"Amy","family":"Sniffen","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]},{"given":"Julien","family":"Blanchet","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]},{"given":"Megan E.","family":"Hillis","sequence":"additional","affiliation":[{"name":"Department of Psychological and Brain Sciences, Dartmouth College"}]},{"given":"Xinyu","family":"Shi","sequence":"additional","affiliation":[{"name":"Department of Informatics, Xiamen University"}]},{"given":"Themistoklis K.","family":"Haris","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]},{"given":"Jason","family":"Liu","sequence":"additional","affiliation":[{"name":"Department of Education, Dartmouth College"}]},{"given":"Jason","family":"Lamberton","sequence":"additional","affiliation":[{"name":"Motion Light Lab, Gallaudet University"}]},{"given":"Melissa","family":"Malzkuhn","sequence":"additional","affiliation":[{"name":"Motion Light Lab, Gallaudet University"}]},{"given":"Lorna C.","family":"Quandt","sequence":"additional","affiliation":[{"name":"Educational Neuroscience Program, Gallaudet University"}]},{"given":"James","family":"Mahoney","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]},{"given":"David J. M.","family":"Kraemer","sequence":"additional","affiliation":[{"name":"Department of Education, Dartmouth College"}]},{"given":"Xia","family":"Zhou","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]},{"given":"Devin","family":"Balkcom","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Dartmouth College"}]}],"member":"320","published-online":{"date-parts":[[2020,12,18]]},"reference":[{"key":"e_1_2_2_1_1","volume-title":"https:\/\/www.lifeprint.com\/asl101\/curriculum\/curriculum.htm. (2019). Online","author":"Lifeprint Cirriculum","year":"2019","unstructured":"2019. Lifeprint Cirriculum | ASL 101. https:\/\/www.lifeprint.com\/asl101\/curriculum\/curriculum.htm. (2019). Online ; accessed 3 November 2019 . 2019. Lifeprint Cirriculum | ASL 101. https:\/\/www.lifeprint.com\/asl101\/curriculum\/curriculum.htm. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_2_1","doi-asserted-by":"crossref","first-page":"175","DOI":"10.1080\/00031305.1992.10475879","article-title":"An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression","volume":"46","author":"Altman N. S.","year":"1992","unstructured":"N. S. Altman . 1992 . An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression . The American Statistician 46 , 3 (1992), 175 -- 185 . http:\/\/www.jstor.org\/stable\/2685209 N. S. Altman. 1992. An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression. The American Statistician 46, 3 (1992), 175--185. http:\/\/www.jstor.org\/stable\/2685209","journal-title":"The American Statistician"},{"key":"e_1_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/2501988.2502045"},{"key":"e_1_2_2_4_1","volume-title":"https:\/autodesk.com\/maya. (2018). Online","author":"Motionbuilder INC.","year":"2019","unstructured":"Autodesk, INC. 2018. Motionbuilder . https:\/autodesk.com\/maya. (2018). Online ; accessed 3 November 2019 . Autodesk, INC. 2018. Motionbuilder. https:\/autodesk.com\/maya. (2018). Online; accessed 3 November 2019."},{"key":"e_1_2_2_5_1","volume-title":"https:\/autodesk.com\/maya. (2019). Online","author":"Maya INC.","year":"2019","unstructured":"Autodesk, INC. 2019. Maya . https:\/autodesk.com\/maya. (2019). Online ; accessed 3 November 2019 . Autodesk, INC. 2019. Maya. https:\/autodesk.com\/maya. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/10.1348"},{"key":"e_1_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377552"},{"key":"e_1_2_2_8_1","volume-title":"American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach. Sensors 18, 10","author":"Chong Teak-Wei","year":"2018","unstructured":"Teak-Wei Chong and Boon-Giin Lee . 2018. American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach. Sensors 18, 10 ( 2018 ). https:\/\/doi.org\/10.3390\/s18103554 10.3390\/s18103554 Teak-Wei Chong and Boon-Giin Lee. 2018. American Sign Language Recognition Using Leap Motion Controller with Machine Learning Approach. Sensors 18, 10 (2018). https:\/\/doi.org\/10.3390\/s18103554"},{"key":"e_1_2_2_9_1","volume-title":"American Sign Language Recognition Using Leap Motion Sensor. In 2014 13th International Conference on Machine Learning and Applications. 541--544","author":"Chuan C.","year":"2014","unstructured":"C. Chuan , E. Regina , and C. Guardino . 2014 . American Sign Language Recognition Using Leap Motion Sensor. In 2014 13th International Conference on Machine Learning and Applications. 541--544 . https:\/\/doi.org\/10.1109\/ICMLA. 2014 .110 10.1109\/ICMLA.2014.110 C. Chuan, E. Regina, and C. Guardino. 2014. American Sign Language Recognition Using Leap Motion Sensor. In 2014 13th International Conference on Machine Learning and Applications. 541--544. https:\/\/doi.org\/10.1109\/ICMLA.2014.110"},{"key":"e_1_2_2_10_1","volume-title":"Signing Gesture Markup Language (SiGML). https:\/\/www.sign-lang.uni-hamburg.de\/hamnosys\/input\/. (2019). Online","year":"2019","unstructured":"DGS-Korpus. 2019. Signing Gesture Markup Language (SiGML). https:\/\/www.sign-lang.uni-hamburg.de\/hamnosys\/input\/. (2019). Online ; accessed 3 November 2019 . DGS-Korpus. 2019. Signing Gesture Markup Language (SiGML). https:\/\/www.sign-lang.uni-hamburg.de\/hamnosys\/input\/. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_11_1","volume-title":"Hand Gesture Recognition with Leap Motion. CoRR abs\/1711.04293","author":"Du Youchen","year":"2017","unstructured":"Youchen Du , Shenglan Liu , Lin Feng , Menghui Chen , and Jie Wu. 2017. Hand Gesture Recognition with Leap Motion. CoRR abs\/1711.04293 ( 2017 ). arXiv:1711.04293 http:\/\/arxiv.org\/abs\/1711.04293 Youchen Du, Shenglan Liu, Lin Feng, Menghui Chen, and Jie Wu. 2017. Hand Gesture Recognition with Leap Motion. CoRR abs\/1711.04293 (2017). arXiv:1711.04293 http:\/\/arxiv.org\/abs\/1711.04293"},{"key":"e_1_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.74"},{"key":"e_1_2_2_13_1","volume-title":"Proc. Intl. Conf. Intelligent Robots and Systems (IROS) 2005 Workshop on Robot Vision for Space Applications.","author":"Fiala Mark","year":"2005","unstructured":"Mark Fiala . 2005 . Artag fiducial marker system applied to vision based spacecraft docking . In Proc. Intl. Conf. Intelligent Robots and Systems (IROS) 2005 Workshop on Robot Vision for Space Applications. Mark Fiala. 2005. Artag fiducial marker system applied to vision based spacecraft docking. In Proc. Intl. Conf. Intelligent Robots and Systems (IROS) 2005 Workshop on Robot Vision for Space Applications."},{"key":"e_1_2_2_14_1","volume-title":"Int. J. Internet Sci. 8 (06","author":"Freelon Deen","year":"2013","unstructured":"Deen Freelon . 2013. ReCal OIR: Ordinal, Interval, and Ratio Intercoder Reliability as a Web Service . Int. J. Internet Sci. 8 (06 2013 ), 10--16. Deen Freelon. 2013. ReCal OIR: Ordinal, Interval, and Ratio Intercoder Reliability as a Web Service. Int. J. Internet Sci. 8 (06 2013), 10--16."},{"key":"e_1_2_2_15_1","volume-title":"The International Scientific Conference eLearning and Software for Education.","author":"Freina Laura","year":"2015","unstructured":"Laura Freina and Michela Ott . 2015 . A literature review on immersive virtual reality in education: state of the art and perspectives . In The International Scientific Conference eLearning and Software for Education. Laura Freina and Michela Ott. 2015. A literature review on immersive virtual reality in education: state of the art and perspectives. In The International Scientific Conference eLearning and Software for Education."},{"key":"e_1_2_2_16_1","volume-title":"Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognition 47 (06","author":"Garrido-Jurado Sergio","year":"2014","unstructured":"Sergio Garrido-Jurado , Rafael Mu\u00f1oz-Salinas , Francisco Madrid-Cuevas , and Manuel Mar\u00edn-Jim\u00e9nez . 2014. Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognition 47 (06 2014 ), 2280--2292. https:\/\/doi.org\/10.1016\/j.patcog.2014.01.005 10.1016\/j.patcog.2014.01.005 Sergio Garrido-Jurado, Rafael Mu\u00f1oz-Salinas, Francisco Madrid-Cuevas, and Manuel Mar\u00edn-Jim\u00e9nez. 2014. Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognition 47 (06 2014), 2280--2292. https:\/\/doi.org\/10.1016\/j.patcog.2014.01.005"},{"key":"e_1_2_2_17_1","volume-title":"https:\/\/www.sign-lang.uni-hamburg.de\/dgs-korpus\/files\/inhalt_pdf\/HamNoSys_Handshapes.pdf. (2019). Online","author":"Hanke Thomas","year":"2019","unstructured":"Thomas Hanke . 2019. Ham NoSys 4 Handshapes Chart . https:\/\/www.sign-lang.uni-hamburg.de\/dgs-korpus\/files\/inhalt_pdf\/HamNoSys_Handshapes.pdf. (2019). Online ; accessed 3 November 2019 . Thomas Hanke. 2019. HamNoSys 4Handshapes Chart. https:\/\/www.sign-lang.uni-hamburg.de\/dgs-korpus\/files\/inhalt_pdf\/HamNoSys_Handshapes.pdf. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_18_1","unstructured":"Thomas Hanke. 2019. HamNoSys-Hamburg Notation System for Sign Languages. https:\/\/www.sign-lang.uni-hamburg.de\/dgs-korpus\/index.php\/hamnosys-97.html. (2019). 2019-11-13 09:12:05 -0500. Thomas Hanke. 2019. HamNoSys-Hamburg Notation System for Sign Languages. https:\/\/www.sign-lang.uni-hamburg.de\/dgs-korpus\/index.php\/hamnosys-97.html. (2019). 2019-11-13 09:12:05 -0500."},{"key":"e_1_2_2_19_1","volume-title":"4th International Conference on Language Resources and Evaluation(LREC).","year":"2004","unstructured":"Hanke, Thomas. 2004 . HamNoSys---Representing sign language data in language resources and language processing contexts . In 4th International Conference on Language Resources and Evaluation(LREC). Hanke, Thomas. 2004. HamNoSys---Representing sign language data in language resources and language processing contexts. In 4th International Conference on Language Resources and Evaluation(LREC)."},{"key":"e_1_2_2_20_1","volume-title":"https:\/\/aslsignbank.haskins.yale.edu\/about\/copyright\/. (2019). Online","author":"Haskin Lab at Yale University","year":"2019","unstructured":"Haskin Lab at Yale University . 2019. ASL Signbank . https:\/\/aslsignbank.haskins.yale.edu\/about\/copyright\/. (2019). Online ; accessed 3 November 2019 . Haskin Lab at Yale University. 2019. ASL Signbank. https:\/\/aslsignbank.haskins.yale.edu\/about\/copyright\/. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_21_1","volume-title":"HTC Vive Pro Is Getting Finger Tracking. (2019). https:\/\/uploadvr.com\/htc-vive-finger-tracking\/ Online","author":"Heaney David","year":"2019","unstructured":"David Heaney . 2019. HTC Vive Pro Is Getting Finger Tracking. (2019). https:\/\/uploadvr.com\/htc-vive-finger-tracking\/ Online ; accessed 3 November 2019 . David Heaney. 2019. HTC Vive Pro Is Getting Finger Tracking. (2019). https:\/\/uploadvr.com\/htc-vive-finger-tracking\/ Online; accessed 3 November 2019."},{"key":"e_1_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3300061.3300117"},{"key":"e_1_2_2_23_1","volume-title":"https:\/\/www.vive.com\/us\/product\/vive-pro\/. (2019). Online","author":"htc Inc. 2019. HTC VIVEPro.","year":"2019","unstructured":"htc Inc. 2019. HTC VIVEPro. https:\/\/www.vive.com\/us\/product\/vive-pro\/. (2019). Online ; accessed 3 November 2019 . htc Inc. 2019. HTC VIVEPro. https:\/\/www.vive.com\/us\/product\/vive-pro\/. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_24_1","volume-title":"Achiou Winter Knit Gloves. (2019). https:\/\/www.amazon.com\/gp\/product\/B077MLRYNN\/ref=ppx_yo_dt_b_asin_title_o02_s00?ie=UTF8&psc=1 Online","author":"Achiou Inc. 2019.","year":"2019","unstructured":"Achiou Inc. 2019. Achiou Winter Knit Gloves. (2019). https:\/\/www.amazon.com\/gp\/product\/B077MLRYNN\/ref=ppx_yo_dt_b_asin_title_o02_s00?ie=UTF8&psc=1 Online ; accessed 3 November 2019 . Achiou Inc. 2019. Achiou Winter Knit Gloves. (2019). https:\/\/www.amazon.com\/gp\/product\/B077MLRYNN\/ref=ppx_yo_dt_b_asin_title_o02_s00?ie=UTF8&psc=1 Online; accessed 3 November 2019."},{"key":"e_1_2_2_25_1","volume-title":"https:\/\/store.arduino.cc\/usa\/arduino-mkr1000 Online","author":"ARDUINO Inc. 2019. MKR1000.","year":"2019","unstructured":"ARDUINO Inc. 2019. MKR1000. ( 2019 ). https:\/\/store.arduino.cc\/usa\/arduino-mkr1000 Online ; accessed 3 November 2019. ARDUINO Inc. 2019. MKR1000. (2019). https:\/\/store.arduino.cc\/usa\/arduino-mkr1000 Online; accessed 3 November 2019."},{"key":"e_1_2_2_26_1","volume-title":"Flex Sensors. (2019). https:\/\/shop.flexpoint.com\/ Online","author":"FlexPoint Inc. 2019.","year":"2019","unstructured":"FlexPoint Inc. 2019. Flex Sensors. (2019). https:\/\/shop.flexpoint.com\/ Online ; accessed 3 November 2019 . FlexPoint Inc. 2019. Flex Sensors. (2019). https:\/\/shop.flexpoint.com\/ Online; accessed 3 November 2019."},{"key":"e_1_2_2_27_1","volume-title":"https:\/\/store.arduino.cc\/usa\/arduino-mkr1000 Online","author":"ValBox Inc. 2019. Cardboard Box.","year":"2019","unstructured":"ValBox Inc. 2019. Cardboard Box. ( 2019 ). https:\/\/store.arduino.cc\/usa\/arduino-mkr1000 Online ; accessed 3 November 2019. ValBox Inc. 2019. Cardboard Box. (2019). https:\/\/store.arduino.cc\/usa\/arduino-mkr1000 Online; accessed 3 November 2019."},{"key":"e_1_2_2_28_1","volume-title":"Force Sensor Datasheet. (2019). https:\/\/cdn2.hubspot.net\/hubfs\/3899023\/Interlinkelectronics%20November2017\/Docs\/Datasheet_FSR.pdf Online","author":"Interlink Electronics Inc. 2019.","year":"2019","unstructured":"Interlink Electronics Inc. 2019. Force Sensor Datasheet. (2019). https:\/\/cdn2.hubspot.net\/hubfs\/3899023\/Interlinkelectronics%20November2017\/Docs\/Datasheet_FSR.pdf Online ; accessed 3 November 2019 . Interlink Electronics Inc. 2019. Force Sensor Datasheet. (2019). https:\/\/cdn2.hubspot.net\/hubfs\/3899023\/Interlinkelectronics%20November2017\/Docs\/Datasheet_FSR.pdf Online; accessed 3 November 2019."},{"key":"e_1_2_2_29_1","volume-title":"Extrinsic flexor muscles generate concurrent flexion of all three finger joints. Journal of biomechanics 35 (01","author":"Kamper Derek","year":"2003","unstructured":"Derek Kamper , T. George Hornby , and William Rymer . 2003. Extrinsic flexor muscles generate concurrent flexion of all three finger joints. Journal of biomechanics 35 (01 2003 ), 1581--9. https:\/\/doi.org\/10.1016\/S0021-9290(02)00229-4 10.1016\/S0021-9290(02)00229-4 Derek Kamper, T. George Hornby, and William Rymer. 2003. Extrinsic flexor muscles generate concurrent flexion of all three finger joints. Journal of biomechanics 35 (01 2003), 1581--9. https:\/\/doi.org\/10.1016\/S0021-9290(02)00229-4"},{"key":"#cr-split#-e_1_2_2_30_1.1","doi-asserted-by":"crossref","unstructured":"Ratchadaporn Kanawong and Aniwat Kanwaratron. 2017. Human Motion Matching for Assisting Standard Thai Folk Dance Learning. 49--53. https:\/\/doi.org\/10.5176\/2251-1679_CGAT17.11 10.5176\/2251-1679_CGAT17.11","DOI":"10.5176\/2251-1679_CGAT17.11"},{"key":"#cr-split#-e_1_2_2_30_1.2","doi-asserted-by":"crossref","unstructured":"Ratchadaporn Kanawong and Aniwat Kanwaratron. 2017. Human Motion Matching for Assisting Standard Thai Folk Dance Learning. 49--53. https:\/\/doi.org\/10.5176\/2251-1679_CGAT17.11","DOI":"10.5176\/2251-1679_CGAT17.11"},{"key":"e_1_2_2_31_1","unstructured":"Bassem Khelil and Hamid Amiri. 2016. Hand Gesture Recognition Using Leap Motion Controller for Recognition of Arabic Sign Language. Bassem Khelil and Hamid Amiri. 2016. Hand Gesture Recognition Using Leap Motion Controller for Recognition of Arabic Sign Language."},{"key":"e_1_2_2_32_1","volume-title":"A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. Journal of Chiropractic Medicine 15 (03","author":"Koo Terry","year":"2016","unstructured":"Terry Koo and Mae Li. 2016. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. Journal of Chiropractic Medicine 15 (03 2016 ). https:\/\/doi.org\/10.1016\/j.jcm.2016.02.012 10.1016\/j.jcm.2016.02.012 Terry Koo and Mae Li. 2016. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. Journal of Chiropractic Medicine 15 (03 2016). https:\/\/doi.org\/10.1016\/j.jcm.2016.02.012"},{"key":"e_1_2_2_33_1","volume-title":"Computing Krippendorff's Alpha-Reliability. (01","year":"2011","unstructured":"klaus krippendorff. 2011. Computing Krippendorff's Alpha-Reliability. (01 2011 ). klaus krippendorff. 2011. Computing Krippendorff's Alpha-Reliability. (01 2011)."},{"key":"e_1_2_2_34_1","volume-title":"12th International Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, 2004. HAPTICS '04. Proceedings.","author":"Lederman S. J.","unstructured":"S. J. Lederman , R. D. Howe , R. L. Klatzky , and C. Hamilton . 2004. Force variability during surface contact with bare finger or rigid probe . In 12th International Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, 2004. HAPTICS '04. Proceedings. S. J. Lederman, R. D. Howe, R. L. Klatzky, and C. Hamilton. 2004. Force variability during surface contact with bare finger or rigid probe. In 12th International Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, 2004. HAPTICS '04. Proceedings."},{"key":"e_1_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2971648.2971738"},{"key":"e_1_2_2_36_1","unstructured":"R. Likert. 1932. A Technique for the Measurement of Attitudes. Number nos. 136-165 in A Technique for the Measurement of Attitudes. publisher not identified. https:\/\/books.google.com\/books?id=9rotAAAAYAAJ R. Likert. 1932. A Technique for the Measurement of Attitudes. Number nos. 136-165 in A Technique for the Measurement of Attitudes. publisher not identified. https:\/\/books.google.com\/books?id=9rotAAAAYAAJ"},{"key":"e_1_2_2_37_1","volume-title":"Leap Motion. (2019). https:\/\/www.leapmotion.com Online","author":"Ltd Ultraleap","year":"2019","unstructured":"Ultraleap Ltd . 2019. Leap Motion. (2019). https:\/\/www.leapmotion.com Online ; accessed 3 November 2019 . Ultraleap Ltd. 2019. Leap Motion. (2019). https:\/\/www.leapmotion.com Online; accessed 3 November 2019."},{"key":"e_1_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3191755"},{"key":"e_1_2_2_39_1","volume-title":"Proc. of the Second International Conference on Information and Communication Technology for Competitive Strategies (ICTCS '16)","author":"Rajesh","unstructured":"Rajesh B. Mapari and Govind Kharat. 2016. American Static Signs Recognition Using Leap Motion Sensor . In Proc. of the Second International Conference on Information and Communication Technology for Competitive Strategies (ICTCS '16) . ACM, New York, NY, USA, Article 67, 5 pages. https:\/\/doi.org\/10.1145\/2905055.2905125 10.1145\/2905055.2905125 Rajesh B. Mapari and Govind Kharat. 2016. American Static Signs Recognition Using Leap Motion Sensor. In Proc. of the Second International Conference on Information and Communication Technology for Competitive Strategies (ICTCS '16). ACM, New York, NY, USA, Article 67, 5 pages. https:\/\/doi.org\/10.1145\/2905055.2905125"},{"key":"e_1_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2014.7025313"},{"key":"e_1_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3134368.3139218"},{"key":"#cr-split#-e_1_2_2_42_1.1","doi-asserted-by":"crossref","unstructured":"Kenneth Mcgraw and S.P. Wong. 1996. Forming Inferences About Some Intraclass Correlation Coefficients. Psychological Methods 1 (03 1996) 30--46. https:\/\/doi.org\/10.1037\/1082-989X.1.1.30 10.1037\/1082-989X.1.1.30","DOI":"10.1037\/1082-989X.1.1.30"},{"key":"#cr-split#-e_1_2_2_42_1.2","doi-asserted-by":"crossref","unstructured":"Kenneth Mcgraw and S.P. Wong. 1996. Forming Inferences About Some Intraclass Correlation Coefficients. Psychological Methods 1 (03 1996) 30--46. https:\/\/doi.org\/10.1037\/1082-989X.1.1.30","DOI":"10.1037\/1082-989X.1.1.30"},{"key":"e_1_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/2632048.2632095"},{"key":"e_1_2_2_44_1","volume-title":"https:\/\/support.xbox.com\/en-US\/xbox-360\/accessories\/kinect-sensor-components Online","year":"2019","unstructured":"Microsoft. 2019. Kinect. ( 2019 ). https:\/\/support.xbox.com\/en-US\/xbox-360\/accessories\/kinect-sensor-components Online ; accessed 3 November 2019. Microsoft. 2019. Kinect. (2019). https:\/\/support.xbox.com\/en-US\/xbox-360\/accessories\/kinect-sensor-components Online; accessed 3 November 2019."},{"key":"e_1_2_2_45_1","volume-title":"http:\/\/www.cs.uu.nl\/docs\/vakken\/mcanim\/mocap-manual\/site\/vicon-blade\/. (2019). Online","author":"Manual Motion Capture","year":"2019","unstructured":"Motion Capture Manual . 2019. Vicon Blade . http:\/\/www.cs.uu.nl\/docs\/vakken\/mcanim\/mocap-manual\/site\/vicon-blade\/. (2019). Online ; accessed 3 November 2019 . Motion Capture Manual. 2019. Vicon Blade. http:\/\/www.cs.uu.nl\/docs\/vakken\/mcanim\/mocap-manual\/site\/vicon-blade\/. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/BSN.2018.8329645"},{"key":"e_1_2_2_47_1","volume-title":"Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 12 (Nov.","author":"Varoquaux Fabian","year":"2011","unstructured":"Pedregosa, Fabian and Varoquaux , Ga\u00ebl and Gramfort , Alexandre and Michel , Vincent and Thirion , Bertrand and Grisel , Olivier and Blondel , Mathieu and Prettenhofer , Peter and Weiss , Ron and Dubourg , Vincent and Vanderplas , Jake and Passos , Alexandre and Cournapeau , David and Brucher , Matthieu and Perrot , Matthieu and Duchesnay , \u00c9douard. 2011 . Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 12 (Nov. 2011), 2825--2830. http:\/\/dl.acm.org\/citation.cfm?id=1953048.2078195 Pedregosa, Fabian and Varoquaux, Ga\u00ebl and Gramfort, Alexandre and Michel, Vincent and Thirion, Bertrand and Grisel, Olivier and Blondel, Mathieu and Prettenhofer, Peter and Weiss, Ron and Dubourg, Vincent and Vanderplas, Jake and Passos, Alexandre and Cournapeau, David and Brucher, Matthieu and Perrot, Matthieu and Duchesnay, \u00c9douard. 2011. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 12 (Nov. 2011), 2825--2830. http:\/\/dl.acm.org\/citation.cfm?id=1953048.2078195"},{"key":"e_1_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/URAI.2016.7734059"},{"key":"e_1_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jocn.2016.09.002"},{"key":"e_1_2_2_50_1","volume-title":"Physix Gear Sport Waterproof Kinesiology Tape. https:\/\/www.amazon.com\/gp\/product\/B017TH9X22\/ref=ppx_yo_dt_b_asin_title_o05_s00?ie=UTF8&psc=1. (2019). Online","author":"PhySix Gear Sport Inc. 2019.","year":"2019","unstructured":"PhySix Gear Sport Inc. 2019. Physix Gear Sport Waterproof Kinesiology Tape. https:\/\/www.amazon.com\/gp\/product\/B017TH9X22\/ref=ppx_yo_dt_b_asin_title_o05_s00?ie=UTF8&psc=1. (2019). Online ; accessed 3 November 2019 . PhySix Gear Sport Inc. 2019. Physix Gear Sport Waterproof Kinesiology Tape. https:\/\/www.amazon.com\/gp\/product\/B017TH9X22\/ref=ppx_yo_dt_b_asin_title_o05_s00?ie=UTF8&psc=1. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3381010"},{"key":"e_1_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097620.3097624"},{"key":"e_1_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIE.2016.2521346"},{"key":"e_1_2_2_54_1","volume-title":"https:\/\/www.stereolabs.com\/zed-mini\/. (2019). Online","author":"Stereolabs Inc. 2019. ZED Mini.","year":"2019","unstructured":"Stereolabs Inc. 2019. ZED Mini. https:\/\/www.stereolabs.com\/zed-mini\/. (2019). Online ; accessed 3 November 2019 . Stereolabs Inc. 2019. ZED Mini. https:\/\/www.stereolabs.com\/zed-mini\/. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_55_1","volume-title":"Development of Master-slave Type Lower Limb Motion Teaching System. In 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS).","author":"Tagami Toshihiro","year":"2018","unstructured":"Toshihiro Tagami , Toshihiro Kawase , Daisuke Morisaki , Ryoken Miyazaki , Tetsuro Miyazaki , Takahiro Kanno , and Kenji Kawashima . 2018 . Development of Master-slave Type Lower Limb Motion Teaching System. In 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). Toshihiro Tagami, Toshihiro Kawase, Daisuke Morisaki, Ryoken Miyazaki, Tetsuro Miyazaki, Takahiro Kanno, and Kenji Kawashima. 2018. Development of Master-slave Type Lower Limb Motion Teaching System. In 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)."},{"key":"e_1_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/WHC.2019.8816118"},{"key":"e_1_2_2_57_1","volume-title":"https:\/\/www.vicon.com Online","author":"Systems Vicon Motion","year":"2019","unstructured":"Vicon Motion Systems Ltd UK. 2019. VICON. ( 2019 ). https:\/\/www.vicon.com Online ; accessed 3 November 2019. Vicon Motion Systems Ltd UK. 2019. VICON. (2019). https:\/\/www.vicon.com Online; accessed 3 November 2019."},{"key":"e_1_2_2_58_1","volume-title":"https:\/\/unity.com\/. (2019). Online","author":"Technologies Unity","year":"2019","unstructured":"Unity Technologies . 2019. Unity Engine . https:\/\/unity.com\/. (2019). Online ; accessed 11 November 2019 . Unity Technologies. 2019. Unity Engine. https:\/\/unity.com\/. (2019). Online; accessed 11 November 2019."},{"key":"e_1_2_2_59_1","volume-title":"Vero Series camera. https:\/\/docs.vicon.com\/display\/Tracker33\/Compatibility+with+Vicon+Vero+cameras. (2019). Online","author":"Systems Vicon Motion","year":"2019","unstructured":"Vicon Motion Systems . 2019. Vero Series camera. https:\/\/docs.vicon.com\/display\/Tracker33\/Compatibility+with+Vicon+Vero+cameras. (2019). Online ; accessed 3 November 2019 . Vicon Motion Systems. 2019. Vero Series camera. https:\/\/docs.vicon.com\/display\/Tracker33\/Compatibility+with+Vicon+Vero+cameras. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_60_1","unstructured":"Vicon Motion Systems Limited. 2006. Vicon MX Hardware System Reference. http:\/\/bdml.stanford.edu\/twiki\/pub\/Haptics\/MotionDisplayKAUST\/ViconHardwareReference.pdf. (2006). Online; accessed 3 November 2019. Vicon Motion Systems Limited. 2006. Vicon MX Hardware System Reference. http:\/\/bdml.stanford.edu\/twiki\/pub\/Haptics\/MotionDisplayKAUST\/ViconHardwareReference.pdf. (2006). Online; accessed 3 November 2019."},{"key":"e_1_2_2_61_1","volume-title":"http:\/\/vhg.cmp.uea.ac.uk\/tech\/jas\/vhg2019\/CWASA-plus-gui-panel.html. (2019). Online","author":"Virtual Humans Group from University of East Anglia. 2019. JASigning.","year":"2019","unstructured":"Virtual Humans Group from University of East Anglia. 2019. JASigning. http:\/\/vhg.cmp.uea.ac.uk\/tech\/jas\/vhg2019\/CWASA-plus-gui-panel.html. (2019). Online ; accessed 3 November 2019 . Virtual Humans Group from University of East Anglia. 2019. JASigning. http:\/\/vhg.cmp.uea.ac.uk\/tech\/jas\/vhg2019\/CWASA-plus-gui-panel.html. (2019). Online; accessed 3 November 2019."},{"key":"e_1_2_2_62_1","volume-title":"Introducing Hand Tracking on Oculus Quest - Bringing Your Real Hands into VR. (2019). https:\/\/www.oculus.com\/blog\/introducing-hand-tracking-on-oculus-quest-bringing-your-real-hands-into-vr\/ Online","author":"Oculus VR.","year":"2019","unstructured":"Oculus VR. 2019. Introducing Hand Tracking on Oculus Quest - Bringing Your Real Hands into VR. (2019). https:\/\/www.oculus.com\/blog\/introducing-hand-tracking-on-oculus-quest-bringing-your-real-hands-into-vr\/ Online ; accessed 11 November 2019 . Oculus VR. 2019. Introducing Hand Tracking on Oculus Quest - Bringing Your Real Hands into VR. (2019). https:\/\/www.oculus.com\/blog\/introducing-hand-tracking-on-oculus-quest-bringing-your-real-hands-into-vr\/ Online; accessed 11 November 2019."},{"key":"e_1_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/1124772.1124941"},{"key":"e_1_2_2_64_1","volume-title":"Microsoft at MWC Barcelona: Introducing Microsoft HoloLens 2. (2019). https:\/\/blogs.microsoft.com\/blog\/2019\/02\/24\/microsoft-at-mwc-barcelona-introducing-microsoft-hololens-2\/ Online","author":"White Julia","year":"2019","unstructured":"Julia White . 2019. Microsoft at MWC Barcelona: Introducing Microsoft HoloLens 2. (2019). https:\/\/blogs.microsoft.com\/blog\/2019\/02\/24\/microsoft-at-mwc-barcelona-introducing-microsoft-hololens-2\/ Online ; accessed 3 November 2019 . Julia White. 2019. Microsoft at MWC Barcelona: Introducing Microsoft HoloLens 2. (2019). https:\/\/blogs.microsoft.com\/blog\/2019\/02\/24\/microsoft-at-mwc-barcelona-introducing-microsoft-hololens-2\/ Online; accessed 3 November 2019."},{"key":"e_1_2_2_65_1","volume-title":"https:\/\/www.lifeprint.com\/asl101\/topics\/highschoolcurriculum.htm. (2020). Online","author":"Vicars Ed.D.","year":"2020","unstructured":"Ed.D. William G. Vicars . 2020. LifePrint. https:\/\/www.lifeprint.com\/asl101\/topics\/highschoolcurriculum.htm. (2020). Online ; accessed 1 August 2020 . Ed.D. William G. Vicars. 2020. LifePrint. https:\/\/www.lifeprint.com\/asl101\/topics\/highschoolcurriculum.htm. (2020). Online; accessed 1 August 2020."},{"key":"e_1_2_2_66_1","volume-title":"Dannenberg","author":"Xia Gus","year":"2018","unstructured":"Gus Xia , Carter Jacobsen , Qianwen Chen , Xingdong Yang , and Roger B . Dannenberg . 2018 . ShIFT: A Semi-haptic Interface for Flute Tutoring. CoRR abs\/1803.06625 (2018). arXiv:1803.06625 http:\/\/arxiv.org\/abs\/1803.06625 Gus Xia, Carter Jacobsen, Qianwen Chen, Xingdong Yang, and Roger B. Dannenberg. 2018. ShIFT: A Semi-haptic Interface for Flute Tutoring. CoRR abs\/1803.06625 (2018). arXiv:1803.06625 http:\/\/arxiv.org\/abs\/1803.06625"},{"key":"e_1_2_2_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2018.2800781"},{"key":"e_1_2_2_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287080"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3432211","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3432211","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3432211","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:47:09Z","timestamp":1750193229000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3432211"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,12,17]]},"references-count":70,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2020,12,17]]}},"alternative-id":["10.1145\/3432211"],"URL":"https:\/\/doi.org\/10.1145\/3432211","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,12,17]]},"assertion":[{"value":"2020-12-18","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}