{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:17:23Z","timestamp":1750220243282,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":62,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,5,12]],"date-time":"2021-05-12T00:00:00Z","timestamp":1620777600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,5,12]]},"DOI":"10.1145\/3490632.3490672","type":"proceedings-article","created":{"date-parts":[[2022,2,25]],"date-time":"2022-02-25T11:25:58Z","timestamp":1645788358000},"page":"158-168","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["Studying Natural User Interfaces for Smart Video Annotation towards Ubiquitous Environments"],"prefix":"10.1145","author":[{"given":"Rui","family":"Rodrigues","sequence":"first","affiliation":[{"name":"Sustain.RD, ESTSetubal Polytechnic Inst. Setubal, Portugal and FCT \/ Nova Lincs, NOVA University of Lisbon, Portugal"}]},{"given":"Rui","family":"Neves Madeira","sequence":"additional","affiliation":[{"name":"Sustain.RD, ESTSetubal Polytechnic Inst. Setubal, Portugal and FCT \/ Nova Lincs, NOVA University of Lisbon, Portugal"}]},{"given":"Nuno","family":"Correia","sequence":"additional","affiliation":[{"name":"FCT \/ Nova Lincs, NOVA University of Lisbon, Portugal"}]}],"member":"320","published-online":{"date-parts":[[2022,2,25]]},"reference":[{"doi-asserted-by":"publisher","key":"e_1_3_2_1_1_1","DOI":"10.5815\/ijmecs.2019.03.03"},{"unstructured":"Amazon Alexa Voice Assistant | Alexa Developer Official Site: 2021. https:\/\/developer.amazon.com\/en-US\/alexa. Accessed: 2021-01-24.  Amazon Alexa Voice Assistant | Alexa Developer Official Site: 2021. https:\/\/developer.amazon.com\/en-US\/alexa. Accessed: 2021-01-24.","key":"e_1_3_2_1_2_1"},{"doi-asserted-by":"crossref","unstructured":"Augstein M. 2019. WeldVUI: Establishing Speech-Based Interfaces in Industrial Applications. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (2019).  Augstein M. 2019. WeldVUI: Establishing Speech-Based Interfaces in Industrial Applications. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (2019).","key":"e_1_3_2_1_3_1","DOI":"10.1007\/978-3-030-29387-1_40"},{"doi-asserted-by":"crossref","unstructured":"Bellegarda J.R. 2014. Spoken Language Understanding for Natural Interaction: The Siri Experience. Natural Interaction with Robots Knowbots and Smartphones.  Bellegarda J.R. 2014. Spoken Language Understanding for Natural Interaction: The Siri Experience. Natural Interaction with Robots Knowbots and Smartphones.","key":"e_1_3_2_1_4_1","DOI":"10.1007\/978-1-4614-8280-2_1"},{"key":"e_1_3_2_1_5_1","volume-title":"Proceedings of the 2011 ACM Multimedia Conference and Workshops","author":"Cabral D.","year":"2011","unstructured":"Cabral , D. 2011 . A creation-tool for contemporary dance using multimodal video annotation. MM\u201911 - Proceedings of the 2011 ACM Multimedia Conference and Workshops (2011). Cabral, D. 2011. A creation-tool for contemporary dance using multimodal video annotation. MM\u201911 - Proceedings of the 2011 ACM Multimedia Conference and Workshops (2011)."},{"doi-asserted-by":"crossref","unstructured":"Cabral D. and Correia N. 2017. Video editing with pen-based technology. Multimedia Tools and Applications. (2017). DOI:https:\/\/doi.org\/10.1007\/s11042-016-3329-y. 10.1007\/s11042-016-3329-y","key":"#cr-split#-e_1_3_2_1_6_1.1","DOI":"10.1007\/s11042-016-3329-y"},{"doi-asserted-by":"crossref","unstructured":"Cabral D. and Correia N. 2017. Video editing with pen-based technology. Multimedia Tools and Applications. (2017). DOI:https:\/\/doi.org\/10.1007\/s11042-016-3329-y.","key":"#cr-split#-e_1_3_2_1_6_1.2","DOI":"10.1007\/s11042-016-3329-y"},{"doi-asserted-by":"crossref","unstructured":"Cambria E. and White B. 2014. Jumping NLP curves: A review of natural language processing research. IEEE Computational Intelligence Magazine.  Cambria E. and White B. 2014. Jumping NLP curves: A review of natural language processing research. IEEE Computational Intelligence Magazine.","key":"e_1_3_2_1_7_1","DOI":"10.1109\/MCI.2014.2307227"},{"key":"e_1_3_2_1_8_1","volume-title":"2001. https:\/\/www.techsmith.com\/video-editor","author":"Camtasia","year":"2021","unstructured":"Camtasia : 2001. https:\/\/www.techsmith.com\/video-editor .html. Accessed : 2021 -06-02. Camtasia: 2001. https:\/\/www.techsmith.com\/video-editor.html. Accessed: 2021-06-02."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_9_1","DOI":"10.1145\/3334480.3382839"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_10_1","DOI":"10.1145\/3290605.3300931"},{"doi-asserted-by":"crossref","unstructured":"Christina C. 2017. Powerpoint Controller using Speech Recognition.  Christina C. 2017. Powerpoint Controller using Speech Recognition.","key":"e_1_3_2_1_11_1","DOI":"10.28932\/jutisi.v3i2.609"},{"doi-asserted-by":"crossref","unstructured":"Cohen P.R. and Oviatt S. 2017. Multimodal speech and pen interfaces. The Handbook of Multimodal-Multisensor Interfaces: Foundations User Modeling and Common Modality Combinations - Volume 1. (2017) 403-447. DOI:https:\/\/doi.org\/10.1145\/3015783.3015795. 10.1145\/3015783.3015795","key":"#cr-split#-e_1_3_2_1_12_1.1","DOI":"10.1145\/3015783.3015795"},{"doi-asserted-by":"crossref","unstructured":"Cohen P.R. and Oviatt S. 2017. Multimodal speech and pen interfaces. The Handbook of Multimodal-Multisensor Interfaces: Foundations User Modeling and Common Modality Combinations - Volume 1. (2017) 403-447. DOI:https:\/\/doi.org\/10.1145\/3015783.3015795.","key":"#cr-split#-e_1_3_2_1_12_1.2","DOI":"10.1145\/3015783.3015795"},{"key":"e_1_3_2_1_13_1","first-page":"4","article-title":"Prepare-participate-connect: Active learning with video annotation","volume":"32","author":"Colasante M.","year":"2016","unstructured":"Colasante , M. and Douglas, K. 2016 . Prepare-participate-connect: Active learning with video annotation . Australasian Journal of Educational Technology. 32 , 4 (Nov. 2016), 68\u201391. DOI:https:\/\/doi.org\/10.14742\/ajet.2123. 10.14742\/ajet.2123 Colasante, M. and Douglas, K. 2016. Prepare-participate-connect: Active learning with video annotation. Australasian Journal of Educational Technology. 32, 4 (Nov. 2016), 68\u201391. DOI:https:\/\/doi.org\/10.14742\/ajet.2123.","journal-title":"Australasian Journal of Educational Technology."},{"unstructured":"Cortana - Your personal productivity assistant: 2021. https:\/\/www.microsoft.com\/en-us\/cortana. Accessed: 2021-01-24.  Cortana - Your personal productivity assistant: 2021. https:\/\/www.microsoft.com\/en-us\/cortana. Accessed: 2021-01-24.","key":"e_1_3_2_1_14_1"},{"unstructured":"Frame.io: 2021. https:\/\/www.frame.io\/. Accessed: 2021-05-25.  Frame.io: 2021. https:\/\/www.frame.io\/. Accessed: 2021-05-25.","key":"e_1_3_2_1_15_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_16_1","DOI":"10.1145\/3379337.3415592"},{"key":"e_1_3_2_1_17_1","first-page":"2","article-title":"50 Years of Progress in Speech and Speaker Recognition Research","volume":"1","author":"Furui S.","year":"2005","unstructured":"Furui , S. 2005 . 50 Years of Progress in Speech and Speaker Recognition Research . ECTI Transactions on Computer and Information Technology (ECTI-CIT). 1 , 2 (Jan. 2005), 64\u201374. DOI:https:\/\/doi.org\/10.37936\/ECTI-CIT.200512.51834. 10.37936\/ECTI-CIT.200512.51834 Furui, S. 2005. 50 Years of Progress in Speech and Speaker Recognition Research. ECTI Transactions on Computer and Information Technology (ECTI-CIT). 1, 2 (Jan. 2005), 64\u201374. DOI:https:\/\/doi.org\/10.37936\/ECTI-CIT.200512.51834.","journal-title":"ECTI Transactions on Computer and Information Technology (ECTI-CIT)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_18_1","DOI":"10.1145\/2807442.2807478"},{"key":"e_1_3_2_1_19_1","volume-title":"ACM International Conference Proceeding Series","author":"Ga\u0161evi\u0107 D.","year":"2014","unstructured":"Ga\u0161evi\u0107 , D. 2014 . Analytics of the effects of video use and instruction to support reflective learning . ACM International Conference Proceeding Series (2014), 123\u2013132. Ga\u0161evi\u0107, D. 2014. Analytics of the effects of video use and instruction to support reflective learning. ACM International Conference Proceeding Series (2014), 123\u2013132."},{"doi-asserted-by":"crossref","unstructured":"Gerard C. and Gerard C. 2021. TensorFlow.js. Practical Machine Learning in JavaScript.  Gerard C. and Gerard C. 2021. TensorFlow.js. Practical Machine Learning in JavaScript.","key":"e_1_3_2_1_20_1","DOI":"10.1007\/978-1-4842-6418-8"},{"volume-title":"Video Research in the Learning Sciences","author":"Goldman R.","unstructured":"Goldman , R. 2014. Video Research in the Learning Sciences . Routledge . Goldman, R. 2014. Video Research in the Learning Sciences. Routledge.","key":"e_1_3_2_1_21_1"},{"unstructured":"Juang B.H. and Rabiner L.R. 2004. Automatic Speech Recognition \u2013 A Brief History of the Technology Development. Elsevier Encyclopedia of Language and Linguistics. (2004).  Juang B.H. and Rabiner L.R. 2004. Automatic Speech Recognition \u2013 A Brief History of the Technology Development. Elsevier Encyclopedia of Language and Linguistics. (2004).","key":"e_1_3_2_1_22_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_23_1","DOI":"10.1145\/3357251.3357581"},{"key":"e_1_3_2_1_24_1","volume-title":"Data Mining, Surveillance, and Authoring. (Aug.","author":"Kipp M.","year":"2012","unstructured":"Kipp , M. 2012. Multimedia Annotation, Querying, and Analysis in Anvil. Multimedia Information Extraction: Advances in Video, Audio, and Imagery Analysis for Search , Data Mining, Surveillance, and Authoring. (Aug. 2012 ), 351\u2013367. DOI:https:\/\/doi.org\/10.1002\/9781118219546.CH21. 10.1002\/9781118219546.CH21 Kipp, M. 2012. Multimedia Annotation, Querying, and Analysis in Anvil. Multimedia Information Extraction: Advances in Video, Audio, and Imagery Analysis for Search, Data Mining, Surveillance, and Authoring. (Aug. 2012), 351\u2013367. DOI:https:\/\/doi.org\/10.1002\/9781118219546.CH21."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_25_1","DOI":"10.1145\/2470654.2481301"},{"key":"e_1_3_2_1_26_1","volume-title":"PARTF","author":"Lemon N.","year":"2013","unstructured":"Lemon , N. 2013. Video annotation for collaborative connections to learning: Case studies from an Australian higher education context. Cutting-Edge Technologies in Higher Education. 6 , PARTF ( 2013 ), 181\u2013214. DOI:https:\/\/doi.org\/10.1108\/S2044-9968(2013)000006F010. 10.1108\/S2044-9968(2013)000006F010 Lemon, N. 2013. Video annotation for collaborative connections to learning: Case studies from an Australian higher education context. Cutting-Edge Technologies in Higher Education. 6, PARTF (2013), 181\u2013214. DOI:https:\/\/doi.org\/10.1108\/S2044-9968(2013)000006F010."},{"key":"e_1_3_2_1_27_1","volume-title":"Alexa vs. Siri vs. Cortana vs. Google Assistant: A Comparison of Speech-Based Natural User Interfaces. Advances in Intelligent Systems and Computing","author":"L\u00f3pez G.","year":"2018","unstructured":"L\u00f3pez , G. 2018. Alexa vs. Siri vs. Cortana vs. Google Assistant: A Comparison of Speech-Based Natural User Interfaces. Advances in Intelligent Systems and Computing ( 2018 ). L\u00f3pez, G. 2018. Alexa vs. Siri vs. Cortana vs. Google Assistant: A Comparison of Speech-Based Natural User Interfaces. Advances in Intelligent Systems and Computing (2018)."},{"doi-asserted-by":"crossref","unstructured":"McShane M. 2017. Natural language understanding (NLU not NLP) in cognitive systems. AI Magazine. (2017). DOI:https:\/\/doi.org\/10.1609\/aimag.v38i4.2745. 10.1609\/aimag.v38i4.2745","key":"#cr-split#-e_1_3_2_1_28_1.1","DOI":"10.1609\/aimag.v38i4.2745"},{"doi-asserted-by":"crossref","unstructured":"McShane M. 2017. Natural language understanding (NLU not NLP) in cognitive systems. AI Magazine. (2017). DOI:https:\/\/doi.org\/10.1609\/aimag.v38i4.2745.","key":"#cr-split#-e_1_3_2_1_28_1.2","DOI":"10.1609\/aimag.v38i4.2745"},{"doi-asserted-by":"crossref","unstructured":"Mehler B. 2016. Multi-modal assessment of on-road demand of voice and manual phone calling and voice navigation entry across two embedded vehicle systems. Ergonomics. (2016). DOI:https:\/\/doi.org\/10.1080\/00140139.2015.1081412. 10.1080\/00140139.2015.1081412","key":"#cr-split#-e_1_3_2_1_29_1.1","DOI":"10.1080\/00140139.2015.1081412"},{"doi-asserted-by":"crossref","unstructured":"Mehler B. 2016. Multi-modal assessment of on-road demand of voice and manual phone calling and voice navigation entry across two embedded vehicle systems. Ergonomics. (2016). DOI:https:\/\/doi.org\/10.1080\/00140139.2015.1081412.","key":"#cr-split#-e_1_3_2_1_29_1.2","DOI":"10.1080\/00140139.2015.1081412"},{"doi-asserted-by":"crossref","unstructured":"Mitrevski M. and Mitrevski M. 2018. Getting Started with Wit.ai. Developing Conversational Interfaces for iOS.  Mitrevski M. and Mitrevski M. 2018. Getting Started with Wit.ai. Developing Conversational Interfaces for iOS.","key":"e_1_3_2_1_30_1","DOI":"10.1007\/978-1-4842-3396-2"},{"unstructured":"ml5js\u00b7Friendly Machine Learning For The Web: 2021. https:\/\/ml5js.org\/. Accessed: 2021-01-25.  ml5js\u00b7Friendly Machine Learning For The Web: 2021. https:\/\/ml5js.org\/. Accessed: 2021-01-25.","key":"e_1_3_2_1_31_1"},{"unstructured":"MotionNotes: 2019. https:\/\/motion-notes.di.fct.unl.pt\/. Accessed: 2020-05-16.  MotionNotes: 2019. https:\/\/motion-notes.di.fct.unl.pt\/. Accessed: 2020-05-16.","key":"e_1_3_2_1_32_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_33_1","DOI":"10.1007\/978-1-4614-4696-5"},{"doi-asserted-by":"crossref","unstructured":"Oviatt S. and Cohen P. 2000. Perceptual user interfaces: multimodal interfaces that process what comes naturally. Communications of the ACM. (2000). DOI:https:\/\/doi.org\/10.1145\/330534.330538. 10.1145\/330534.330538","key":"#cr-split#-e_1_3_2_1_34_1.1","DOI":"10.1145\/330534.330538"},{"doi-asserted-by":"crossref","unstructured":"Oviatt S. and Cohen P. 2000. Perceptual user interfaces: multimodal interfaces that process what comes naturally. Communications of the ACM. (2000). DOI:https:\/\/doi.org\/10.1145\/330534.330538.","key":"#cr-split#-e_1_3_2_1_34_1.2","DOI":"10.1145\/330534.330538"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_35_1","DOI":"10.1145\/2723576.2723611"},{"doi-asserted-by":"crossref","unstructured":"Qattous H. 2016. Teachme A Gesture Recognition System with Customization Feature.  Qattous H. 2016. Teachme A Gesture Recognition System with Customization Feature.","key":"e_1_3_2_1_36_1","DOI":"10.14569\/IJACSA.2016.071107"},{"unstructured":"Radziwill N. and Benton M. 2017. Evaluating quality of chatbots and intelligent conversational agents. arXiv.  Radziwill N. and Benton M. 2017. Evaluating quality of chatbots and intelligent conversational agents. arXiv.","key":"e_1_3_2_1_37_1"},{"doi-asserted-by":"crossref","unstructured":"El Raheb K. 2017. BalOnSe: Temporal aspects of dance movement and its ontological representation. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (2017).  El Raheb K. 2017. BalOnSe: Temporal aspects of dance movement and its ontological representation. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (2017).","key":"e_1_3_2_1_38_1","DOI":"10.1007\/978-3-319-58451-5_4"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_39_1","DOI":"10.1145\/2948910.2948961"},{"doi-asserted-by":"crossref","unstructured":"Rich P.J. and Hannafin M. 2009. Video Annotation Tools Technologies to Scaffold Structure and Transform Teacher Reflection. (2009). DOI:https:\/\/doi.org\/10.1177\/0022487108328486. 10.1177\/0022487108328486","key":"#cr-split#-e_1_3_2_1_40_1.1","DOI":"10.1177\/0022487108328486"},{"doi-asserted-by":"crossref","unstructured":"Rich P.J. and Hannafin M. 2009. Video Annotation Tools Technologies to Scaffold Structure and Transform Teacher Reflection. (2009). DOI:https:\/\/doi.org\/10.1177\/0022487108328486.","key":"#cr-split#-e_1_3_2_1_40_1.2","DOI":"10.1177\/0022487108328486"},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_41_1","DOI":"10.1109\/TLT.2012.15"},{"doi-asserted-by":"crossref","unstructured":"Rodrigues R. 2019. Multimodal Web Based Video Annotator with Real-Time Human Pose Estimation. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 11872 LNCS (2019) 23-30. DOI:https:\/\/doi.org\/10.1007\/978-3-030-33617-2_3. 10.1007\/978-3-030-33617-2_3","key":"#cr-split#-e_1_3_2_1_42_1.1","DOI":"10.1007\/978-3-030-33617-2_3"},{"doi-asserted-by":"crossref","unstructured":"Rodrigues R. 2019. Multimodal Web Based Video Annotator with Real-Time Human Pose Estimation. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 11872 LNCS (2019) 23-30. DOI:https:\/\/doi.org\/10.1007\/978-3-030-33617-2_3.","key":"#cr-split#-e_1_3_2_1_42_1.2","DOI":"10.1007\/978-3-030-33617-2_3"},{"doi-asserted-by":"crossref","unstructured":"Singh A.K. 2020. Voice Controlled Media Player: A Use Case to Demonstrate an On-premise Speech Command Recognition System. Communications in Computer and Information Science. 1209 CCIS (2020) 186-197. DOI:https:\/\/doi.org\/10.1007\/978-981-15-4828-4_16. 10.1007\/978-981-15-4828-4_16","key":"#cr-split#-e_1_3_2_1_43_1.1","DOI":"10.1007\/978-981-15-4828-4_16"},{"doi-asserted-by":"crossref","unstructured":"Singh A.K. 2020. Voice Controlled Media Player: A Use Case to Demonstrate an On-premise Speech Command Recognition System. Communications in Computer and Information Science. 1209 CCIS (2020) 186-197. DOI:https:\/\/doi.org\/10.1007\/978-981-15-4828-4_16.","key":"#cr-split#-e_1_3_2_1_43_1.2","DOI":"10.1007\/978-981-15-4828-4_16"},{"key":"e_1_3_2_1_44_1","volume-title":"The choreographer's notebook-a video annotation system for dancers and choreographers. C and C 2011 - Proceedings of the 8th ACM Conference on Creativity and Cognition","author":"Singh V.","year":"2011","unstructured":"Singh , V. 2011. The choreographer's notebook-a video annotation system for dancers and choreographers. C and C 2011 - Proceedings of the 8th ACM Conference on Creativity and Cognition ( 2011 ). Singh, V. 2011. The choreographer's notebook-a video annotation system for dancers and choreographers. C and C 2011 - Proceedings of the 8th ACM Conference on Creativity and Cognition (2011)."},{"unstructured":"Siri - Apple: 2021. https:\/\/www.apple.com\/siri\/. Accessed: 2021-01-24.  Siri - Apple: 2021. https:\/\/www.apple.com\/siri\/. Accessed: 2021-01-24.","key":"e_1_3_2_1_45_1"},{"unstructured":"Smilkov D. 2019. Tensorflow.JS: Machine learning for the web and beyond. arXiv.  Smilkov D. 2019. Tensorflow.JS: Machine learning for the web and beyond. arXiv.","key":"e_1_3_2_1_46_1"},{"key":"e_1_3_2_1_47_1","first-page":"1","article-title":"The effect of multimedia use on the teaching and learning of Social Sciences at tertiary level: a case study","volume":"17","year":"2017","unstructured":"DE SOUSA, L. 2017 . The effect of multimedia use on the teaching and learning of Social Sciences at tertiary level: a case study . Yesterday and Today. 17 (2017), 1 \u2013 22 . DOI:https:\/\/doi.org\/10.17159\/2223-0386\/2017\/n17a1. 10.17159\/2223-0386 DE SOUSA, L. 2017. The effect of multimedia use on the teaching and learning of Social Sciences at tertiary level: a case study. Yesterday and Today. 17 (2017), 1\u201322. DOI:https:\/\/doi.org\/10.17159\/2223-0386\/2017\/n17a1.","journal-title":"Yesterday and Today."},{"doi-asserted-by":"publisher","key":"e_1_3_2_1_48_1","DOI":"10.1109\/TVCG.2017.2745219"},{"doi-asserted-by":"crossref","unstructured":"Stevens R. 2002. VideoTraces. (2002).  Stevens R. 2002. VideoTraces. (2002).","key":"e_1_3_2_1_49_1","DOI":"10.3115\/1658616.1658702"},{"key":"e_1_3_2_1_50_1","volume-title":"Asia Pacific Journal of Educators and Education. 32","author":"Tang D.","year":"2017","unstructured":"Tang , D. 2017 . EFFECTIVENESS OF AUDIO-VISUAL AIDS IN TEACHING LOWER SECONDARY SCIENCE IN A RURAL SECONDARY SCHOOL . Asia Pacific Journal of Educators and Education. 32 , (2017), 91\u2013106. DOI:https:\/\/doi.org\/10.21315\/apjee2017.32.7. 10.21315\/apjee2017.32.7 Tang, D. 2017. EFFECTIVENESS OF AUDIO-VISUAL AIDS IN TEACHING LOWER SECONDARY SCIENCE IN A RURAL SECONDARY SCHOOL. Asia Pacific Journal of Educators and Education. 32, (2017), 91\u2013106. DOI:https:\/\/doi.org\/10.21315\/apjee2017.32.7."},{"doi-asserted-by":"crossref","unstructured":"Turk M. 2014. Multimodal interaction: A review. Pattern Recognition Letters.  Turk M. 2014. Multimodal interaction: A review. Pattern Recognition Letters.","key":"e_1_3_2_1_51_1","DOI":"10.1016\/j.patrec.2013.07.003"},{"unstructured":"Vimeo: 2021. https:\/\/vimeo.com\/features\/video-collaboration. Accessed: 2021-06-05.  Vimeo: 2021. https:\/\/vimeo.com\/features\/video-collaboration. Accessed: 2021-06-05.","key":"e_1_3_2_1_52_1"},{"unstructured":"Wipster | Review Software: 2021. https:\/\/wipster.io\/. Accessed: 2021-06-15.  Wipster | Review Software: 2021. https:\/\/wipster.io\/. Accessed: 2021-06-15.","key":"e_1_3_2_1_53_1"},{"key":"e_1_3_2_1_54_1","volume-title":"Proceedings of the 5th International Conference on Language Resources and Evaluation, LREC 2006","author":"Wittenburg P.","year":"2006","unstructured":"Wittenburg , P. 2006 . ELAN: A professional framework for multimodality research . Proceedings of the 5th International Conference on Language Resources and Evaluation, LREC 2006 (2006). Wittenburg, P. 2006. ELAN: A professional framework for multimodality research. Proceedings of the 5th International Conference on Language Resources and Evaluation, LREC 2006 (2006)."}],"event":{"acronym":"MUM 2021","name":"MUM 2021: 20th International Conference on Mobile and Ubiquitous Multimedia","location":"Leuven Belgium"},"container-title":["Proceedings of the 20th International Conference on Mobile and Ubiquitous Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3490632.3490672","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3490632.3490672","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:29Z","timestamp":1750188629000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3490632.3490672"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,5,12]]},"references-count":62,"alternative-id":["10.1145\/3490632.3490672","10.1145\/3490632"],"URL":"https:\/\/doi.org\/10.1145\/3490632.3490672","relation":{},"subject":[],"published":{"date-parts":[[2021,5,12]]},"assertion":[{"value":"2022-02-25","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}