{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T18:20:31Z","timestamp":1776882031037,"version":"3.51.2"},"publisher-location":"New York, NY, USA","reference-count":72,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,18]],"date-time":"2023-06-18T00:00:00Z","timestamp":1687046400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["112778"],"award-info":[{"award-number":["112778"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["153397"],"award-info":[{"award-number":["153397"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Commonwealth Cyber Initiative"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,18]]},"DOI":"10.1145\/3581791.3596830","type":"proceedings-article","created":{"date-parts":[[2023,6,16]],"date-time":"2023-06-16T17:52:21Z","timestamp":1686937941000},"page":"370-382","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":30,"title":["BystandAR: Protecting Bystander Visual Data in Augmented Reality Systems"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1315-5066","authenticated-orcid":false,"given":"Matthew","family":"Corbett","sequence":"first","affiliation":[{"name":"Virginia Tech, Blacksburg, VA, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3292-1130","authenticated-orcid":false,"given":"Brendan","family":"David-John","sequence":"additional","affiliation":[{"name":"Virginia Tech, Blacksburg, VA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3695-0991","authenticated-orcid":false,"given":"Jiacheng","family":"Shang","sequence":"additional","affiliation":[{"name":"Montclair State University, Montclair, NJ, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1136-9909","authenticated-orcid":false,"given":"Y. Charlie","family":"Hu","sequence":"additional","affiliation":[{"name":"Purdue University, West Lafeyette, IN, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0149-7509","authenticated-orcid":false,"given":"Bo","family":"Ji","sequence":"additional","affiliation":[{"name":"Virginia Tech, Blacksburg, VA, United States of America"}]}],"member":"320","published-online":{"date-parts":[[2023,6,18]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/2906388.2906412"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0059312"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3506857"},{"key":"e_1_3_2_1_4_1","volume-title":"Global Mobile Augmented Reality (AR) user devices","author":"Alsop Thomas","year":"2024","unstructured":"Thomas Alsop . 2022. Global Mobile Augmented Reality (AR) user devices 2024 . https:\/\/www.statista.com\/statistics\/1098630\/global-mobile-augmented-reality-ar-users\/ Thomas Alsop. 2022. Global Mobile Augmented Reality (AR) user devices 2024. https:\/\/www.statista.com\/statistics\/1098630\/global-mobile-augmented-reality-ar-users\/"},{"key":"e_1_3_2_1_5_1","volume-title":"The Psychology of Interpersonal Behaviour","author":"Argyle M.","unstructured":"M. Argyle . 1994. The Psychology of Interpersonal Behaviour . Penguin Books Limited . https:\/\/books.google.com\/books?id=VQOzdOxJFZAC M. Argyle. 1994. The Psychology of Interpersonal Behaviour. Penguin Books Limited. https:\/\/books.google.com\/books?id=VQOzdOxJFZAC"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1098\/rsos.160086"},{"key":"e_1_3_2_1_7_1","volume-title":"Depth Information in Human Gait Analysis: An Experimental Study on Gender Recognition","author":"Borr\u00e0s Ricard","unstructured":"Ricard Borr\u00e0s , \u00c0gata Lapedriza , and Laura Igual . 2012. Depth Information in Human Gait Analysis: An Experimental Study on Gender Recognition . In Image Analysis and Recognition, Aur\u00e9lio Campilho and Mohamed Kamel (Eds.). Springer Berlin Heidelberg , Berlin, Heidelberg , 98--105. Ricard Borr\u00e0s, \u00c0gata Lapedriza, and Laura Igual. 2012. Depth Information in Human Gait Analysis: An Experimental Study on Gender Recognition. In Image Analysis and Recognition, Aur\u00e9lio Campilho and Mohamed Kamel (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 98--105."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSAI.2017.8248353"},{"key":"e_1_3_2_1_9_1","volume-title":"Privacy-Preserving Action Recognition for Smart Hospitals using Low-Resolution Depth Images. CoRR abs\/1811.09950","author":"Chou Edward","year":"2018","unstructured":"Edward Chou , Matthew Tan , Cherry Zou , Michelle Guo , Albert Haque , Arnold Milstein , and Li Fei-Fei . 2018. Privacy-Preserving Action Recognition for Smart Hospitals using Low-Resolution Depth Images. CoRR abs\/1811.09950 ( 2018 ). arXiv:1811.09950 Edward Chou, Matthew Tan, Cherry Zou, Michelle Guo, Albert Haque, Arnold Milstein, and Li Fei-Fei. 2018. Privacy-Preserving Action Recognition for Smart Hospitals using Low-Resolution Depth Images. CoRR abs\/1811.09950 (2018). arXiv:1811.09950"},{"key":"e_1_3_2_1_10_1","volume-title":"Facial recognition beats the Covid-mask challenge. BBC. Retrieved","author":"Clayton James","year":"2022","unstructured":"James Clayton . 2022. Facial recognition beats the Covid-mask challenge. BBC. Retrieved Nov 6, 2022 from https:\/\/www.bbc.com\/news\/technology-56517033 James Clayton. 2022. Facial recognition beats the Covid-mask challenge. BBC. Retrieved Nov 6, 2022 from https:\/\/www.bbc.com\/news\/technology-56517033"},{"key":"e_1_3_2_1_11_1","unstructured":"David Darling. 2021. Automated Privacy Protection for Mobile Device Users and Bystanders in Public Spaces. Master's thesis. Retrieved from https:\/\/scholarworks.uark.edu\/etd\/4218.  David Darling. 2021. Automated Privacy Protection for Mobile Device Users and Bystanders in Public Spaces. Master's thesis. Retrieved from https:\/\/scholarworks.uark.edu\/etd\/4218."},{"key":"e_1_3_2_1_12_1","first-page":"1","article-title":"Identification of Subjects and Bystanders in Photos with Feature-Based Machine Learning. In IEEE INFOCOM 2019 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS). IEEE, 445 Hoes Lane, Piscataway","volume":"08854","author":"Darling David","year":"2019","unstructured":"David Darling , Ang Li , and Qinghua Li . 2019 . Identification of Subjects and Bystanders in Photos with Feature-Based Machine Learning. In IEEE INFOCOM 2019 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS). IEEE, 445 Hoes Lane, Piscataway , NJ 08854. , 1 -- 6 . David Darling, Ang Li, and Qinghua Li. 2019. Identification of Subjects and Bystanders in Photos with Feature-Based Machine Learning. In IEEE INFOCOM 2019 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS). IEEE, 445 Hoes Lane, Piscataway, NJ 08854., 1--6.","journal-title":"NJ"},{"key":"e_1_3_2_1_13_1","volume-title":"Automated Bystander Detection and Anonymization in Mobile Photography. In International Conference on Security and Privacy in Communication Systems. Springer, 402--424","author":"Darling David","year":"2020","unstructured":"David Darling , Ang Li , and Qinghua Li . 2020 . Automated Bystander Detection and Anonymization in Mobile Photography. In International Conference on Security and Privacy in Communication Systems. Springer, 402--424 . David Darling, Ang Li, and Qinghua Li. 2020. Automated Bystander Detection and Anonymization in Mobile Photography. In International Conference on Security and Privacy in Communication Systems. Springer, 402--424."},{"key":"e_1_3_2_1_14_1","volume-title":"SOUPS 2021 Workshop.","author":"David-John Brendan","year":"2021","unstructured":"Brendan David-John , Diane Hosfelt , Kevin Butler , and Eakta Jain . 2021 . Let's SOUP up XR: Collected thoughts from an IEEE VR workshop on privacy in mixed reality. In VR4Sec: Security for VR and VR for Security , SOUPS 2021 Workshop. Brendan David-John, Diane Hosfelt, Kevin Butler, and Eakta Jain. 2021. Let's SOUP up XR: Collected thoughts from an IEEE VR workshop on privacy in mixed reality. In VR4Sec: Security for VR and VR for Security, SOUPS 2021 Workshop."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359626"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2021.616471"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2556288.2557352"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3161190"},{"key":"e_1_3_2_1_19_1","volume-title":"Microsoft. Retrieved","year":"2021","unstructured":"Drewbatgit. 2021 . App Capability declarations - UWP applications . Microsoft. Retrieved Sept 29, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/uwp\/packaging\/app-capability-declarations Drewbatgit. 2021. App Capability declarations - UWP applications. Microsoft. Retrieved Sept 29, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/uwp\/packaging\/app-capability-declarations"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581565"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2335356.2335360"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICACITE53722.2022.9823786"},{"key":"e_1_3_2_1_23_1","volume-title":"Google. Retrieved","year":"2023","unstructured":"Google. 2023 . Why On-Device Machine Learning ? Google. Retrieved Oct 15, 2022 from https:\/\/developers.google.com\/learn\/topics\/on-device-ml\/learn-more Google. 2023. Why On-Device Machine Learning? Google. Retrieved Oct 15, 2022 from https:\/\/developers.google.com\/learn\/topics\/on-device-ml\/learn-more"},{"key":"e_1_3_2_1_24_1","volume-title":"Realtime 3D Object Detection for Headsets. CoRR abs\/2201.08812","author":"Guan Yongjie","year":"2022","unstructured":"Yongjie Guan , Xueyu Hou , Nan Wu , Bo Han , and Tao Han . 2022. Realtime 3D Object Detection for Headsets. CoRR abs\/2201.08812 ( 2022 ). arXiv:2201.08812 Yongjie Guan, Xueyu Hou, Nan Wu, Bo Han, and Tao Han. 2022. Realtime 3D Object Detection for Headsets. CoRR abs\/2201.08812 (2022). arXiv:2201.08812"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3463914.3463918"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP40000.2020.00097"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458864.3467676"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1155\/2016\/8087545"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2013.31"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3498361.3538932"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3240167.3240174"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/2785830.2785842"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Pavan Kunchala. 2021. Real-time age gender detection using opencv. https:\/\/medium.com\/analytics-vidhya\/real-time-age-gender-detection-using-opencv-fa705fe0e1fa  Pavan Kunchala. 2021. Real-time age gender detection using opencv. https:\/\/medium.com\/analytics-vidhya\/real-time-age-gender-detection-using-opencv-fa705fe0e1fa","DOI":"10.1299\/jsmermd.2021.1P2-G01"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3524020"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/SAHCN.2016.7733008"},{"key":"e_1_3_2_1_36_1","volume-title":"GitHub. Retrieved","year":"2022","unstructured":"Linzaer. 2022 . Ultra-Light-Fast-Generic-Face-Detector-1MB . GitHub. Retrieved Oct 15, 2022 from https:\/\/github.com\/Linzaer\/Ultra-Light-Fast-Generic-Face-Detector-1MB Linzaer. 2022. Ultra-Light-Fast-Generic-Face-Detector-1MB. GitHub. Retrieved Oct 15, 2022 from https:\/\/github.com\/Linzaer\/Ultra-Light-Fast-Generic-Face-Detector-1MB"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3300061.3300116"},{"key":"e_1_3_2_1_38_1","volume-title":"MagicLeap. Retrieved","year":"2022","unstructured":"MagicLeap. 2022 . Eye Tracking - Unity . MagicLeap. Retrieved Oct 16, 2022 from https:\/\/ml1-developer.magicleap.com\/en-us\/learn\/guides\/eye-tracking-tutorial-unity MagicLeap. 2022. Eye Tracking - Unity. MagicLeap. Retrieved Oct 16, 2022 from https:\/\/ml1-developer.magicleap.com\/en-us\/learn\/guides\/eye-tracking-tutorial-unity"},{"key":"e_1_3_2_1_39_1","volume-title":"AN ADVANCED VR DEVICE FOR COLLABORATION AND CREATION. Meta. Retrieved","year":"2022","unstructured":"Meta. 2022 . INTRODUCING META QUEST PRO , AN ADVANCED VR DEVICE FOR COLLABORATION AND CREATION. Meta. Retrieved Oct 16, 2022 from https:\/\/www.oculus.com\/blog\/meta-quest-pro-price-release-date\/ Meta. 2022. INTRODUCING META QUEST PRO, AN ADVANCED VR DEVICE FOR COLLABORATION AND CREATION. Meta. Retrieved Oct 16, 2022 from https:\/\/www.oculus.com\/blog\/meta-quest-pro-price-release-date\/"},{"key":"e_1_3_2_1_40_1","volume-title":"Microsoft. Retrieved","year":"2021","unstructured":"Microsoft. 2021 . Hololens 2 . Microsoft. Retrieved Oct 15, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/develop\/advanced-concepts\/hologram-stability Microsoft. 2021. Hololens 2. Microsoft. Retrieved Oct 15, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/develop\/advanced-concepts\/hologram-stability"},{"key":"e_1_3_2_1_41_1","volume-title":"Azure Spatial Anchors. Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022. Azure Spatial Anchors. Microsoft. Retrieved Nov 6, 2022 from https:\/\/azure.microsoft.com\/en-us\/products\/spatial-anchors\/ Microsoft. 2022. Azure Spatial Anchors. Microsoft. Retrieved Nov 6, 2022 from https:\/\/azure.microsoft.com\/en-us\/products\/spatial-anchors\/"},{"key":"e_1_3_2_1_42_1","volume-title":"Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022 . Eye tracking on HoloLens 2 . Microsoft. Retrieved Oct 16, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/design\/eye-tracking Microsoft. 2022. Eye tracking on HoloLens 2. Microsoft. Retrieved Oct 16, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/design\/eye-tracking"},{"key":"e_1_3_2_1_43_1","volume-title":"Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022 . FaceDetector Class . Microsoft. Retrieved Oct 16, 2022 from https:\/\/learn.microsoft.com\/en-us\/uwp\/api\/windows.media.faceanalysis.facedetector?view=winrt-22621 Microsoft. 2022. FaceDetector Class. Microsoft. Retrieved Oct 16, 2022 from https:\/\/learn.microsoft.com\/en-us\/uwp\/api\/windows.media.faceanalysis.facedetector?view=winrt-22621"},{"key":"e_1_3_2_1_44_1","volume-title":"Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022 . Hologram stability . Microsoft. Retrieved Oct 15, 2022 from https:\/\/www.microsoft.com\/en-us\/hololens\/hardware Microsoft. 2022. Hologram stability. Microsoft. Retrieved Oct 15, 2022 from https:\/\/www.microsoft.com\/en-us\/hololens\/hardware"},{"key":"e_1_3_2_1_45_1","volume-title":"Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022 . Holographic face tracking sample . Microsoft. Retrieved Oct 15, 2022 from https:\/\/learn.microsoft.com\/en-us\/samples\/microsoft\/windows-universal-samples\/holographicfacetracking\/ Microsoft. 2022. Holographic face tracking sample. Microsoft. Retrieved Oct 15, 2022 from https:\/\/learn.microsoft.com\/en-us\/samples\/microsoft\/windows-universal-samples\/holographicfacetracking\/"},{"key":"e_1_3_2_1_46_1","volume-title":"Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022 . MediaCapture Class . Microsoft. Retrieved Oct 16, 2022 from https:\/\/learn.microsoft.com\/en-us\/uwp\/api\/windows.media.capture.mediacapture?view=winrt-22621 Microsoft. 2022. MediaCapture Class. Microsoft. Retrieved Oct 16, 2022 from https:\/\/learn.microsoft.com\/en-us\/uwp\/api\/windows.media.capture.mediacapture?view=winrt-22621"},{"key":"e_1_3_2_1_47_1","volume-title":"Mixed reality capture overview. Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022. Mixed reality capture overview. Microsoft. Retrieved Nov 6, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/develop\/advanced-concepts\/mixed-reality-capture-overview Microsoft. 2022. Mixed reality capture overview. Microsoft. Retrieved Nov 6, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/develop\/advanced-concepts\/mixed-reality-capture-overview"},{"key":"e_1_3_2_1_48_1","first-page":"2022","volume-title":"Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022 . What is Mixed Reality Toolkit 2 ? Microsoft. Retrieved Oct 13, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/mrtk-unity\/mrtk2\/?view=mrtkunity- 2022 - 2005 Microsoft. 2022. What is Mixed Reality Toolkit 2? Microsoft. Retrieved Oct 13, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/mrtk-unity\/mrtk2\/?view=mrtkunity-2022-05"},{"key":"e_1_3_2_1_49_1","volume-title":"World locking and spatial anchors in Unity. Microsoft. Retrieved","year":"2022","unstructured":"Microsoft. 2022. World locking and spatial anchors in Unity. Microsoft. Retrieved Nov 6, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/develop\/unity\/spatial-anchors-in-unity?tabs=wlt Microsoft. 2022. World locking and spatial anchors in Unity. Microsoft. Retrieved Nov 6, 2022 from https:\/\/learn.microsoft.com\/en-us\/windows\/mixed-reality\/develop\/unity\/spatial-anchors-in-unity?tabs=wlt"},{"key":"e_1_3_2_1_50_1","volume-title":"Introduction to Motion Estimation with Optical Flow. NanoNets. Retrieved","year":"2022","unstructured":"NanoNets. 2022. Introduction to Motion Estimation with Optical Flow. NanoNets. Retrieved Dec 8, 2022 from https:\/\/nanonets.com\/blog\/optical-flow\/ NanoNets. 2022. Introduction to Motion Estimation with Optical Flow. NanoNets. Retrieved Dec 8, 2022 from https:\/\/nanonets.com\/blog\/optical-flow\/"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/358916.358991"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/MITP.2017.42"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.3390\/iot1020013"},{"key":"e_1_3_2_1_54_1","volume-title":"The Ease-of-use of Unity's Networking with the Performance & Reliability of Photon Realtime. PUN. Retrieved","year":"2022","unstructured":"Photon. 2022. The Ease-of-use of Unity's Networking with the Performance & Reliability of Photon Realtime. PUN. Retrieved Nov 6, 2022 from https:\/\/www.photonengine.com\/pun Photon. 2022. The Ease-of-use of Unity's Networking with the Performance & Reliability of Photon Realtime. PUN. Retrieved Nov 6, 2022 from https:\/\/www.photonengine.com\/pun"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2021.3062546"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1108\/ILS-11-2020-0236"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/2638728.2641707"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/2660267.2660319"},{"key":"e_1_3_2_1_59_1","unstructured":"Joe Saballa. 2022. US Army OKS acquisition of 5 000 IVAS goggles after year-long delay. https:\/\/www.thedefensepost.com\/2022\/09\/05\/us-army-ivas-goggles-2\/  Joe Saballa. 2022. US Army OKS acquisition of 5 000 IVAS goggles after year-long delay. https:\/\/www.thedefensepost.com\/2022\/09\/05\/us-army-ivas-goggles-2\/"},{"key":"e_1_3_2_1_60_1","volume-title":"Eye contact: An introduction to its role in communication","author":"Schulz Jodi","unstructured":"Jodi Schulz . 20212. Eye contact: An introduction to its role in communication . Michigan State University Extension . Jodi Schulz. 20212. Eye contact: An introduction to its role in communication. Michigan State University Extension."},{"key":"e_1_3_2_1_61_1","volume-title":"Deep Face Detection with MTCNN in Python. None. Retrieved","author":"Serengil Sefik Ilkin","year":"2022","unstructured":"Sefik Ilkin Serengil . 2022. Deep Face Detection with MTCNN in Python. None. Retrieved Dec 8, 2022 from sefiks.com\/2020\/09\/09\/deep-face-detection-with-mtcnn-in-python\/ Sefik Ilkin Serengil. 2022. Deep Face Detection with MTCNN in Python. None. Retrieved Dec 8, 2022 from sefiks.com\/2020\/09\/09\/deep-face-detection-with-mtcnn-in-python\/"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICEET53442.2021.9659697"},{"key":"e_1_3_2_1_63_1","unstructured":"Jeff Shepard. 2022. What sensors are used in AR\/VR systems? https:\/\/www.sensortips.com\/featured\/what-sensors-are-used-in-ar-vr-systems-faq\/  Jeff Shepard. 2022. What sensors are used in AR\/VR systems? https:\/\/www.sensortips.com\/featured\/what-sensors-are-used-in-ar-vr-systems-faq\/"},{"key":"e_1_3_2_1_64_1","volume-title":"Cardea: Context-Aware Visual Privacy Protection from Pervasive Cameras. CoRR abs\/1610.00889","author":"Shu Jiayu","year":"2016","unstructured":"Jiayu Shu , Rui Zheng , and Pan Hui . 2016 . Cardea: Context-Aware Visual Privacy Protection from Pervasive Cameras. CoRR abs\/1610.00889 (2016). arXiv:1610.00889 Jiayu Shu, Rui Zheng, and Pan Hui. 2016. Cardea: Context-Aware Visual Privacy Protection from Pervasive Cameras. CoRR abs\/1610.00889 (2016). arXiv:1610.00889"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/3314111.3319913"},{"key":"e_1_3_2_1_66_1","volume-title":"Retrieved","year":"2022","unstructured":"Techjury. 2022 . 29+ augmented reality stats to keep you sharp in 2022 . Retrieved Sept 29, 2022 from https:\/\/techjury.net\/blog\/augmented-reality-stats\/ Techjury. 2022. 29+ augmented reality stats to keep you sharp in 2022. Retrieved Sept 29, 2022 from https:\/\/techjury.net\/blog\/augmented-reality-stats\/"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1007\/11551201_5"},{"key":"e_1_3_2_1_68_1","volume-title":"Unity. Retrieved","year":"2022","unstructured":"Unity. 2022 . GameObject . Unity. Retrieved Oct 16, 2022 from https:\/\/docs.unity3d.com\/ScriptReference\/GameObject.html Unity. 2022. GameObject. Unity. Retrieved Oct 16, 2022 from https:\/\/docs.unity3d.com\/ScriptReference\/GameObject.html"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1109\/SMARTCOMP52413.2021.00021"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/365024.365119"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.2990354"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3517260"}],"event":{"name":"MobiSys '23: 21st Annual International Conference on Mobile Systems, Applications and Services","location":"Helsinki Finland","acronym":"MobiSys '23","sponsor":["SIGMOBILE ACM Special Interest Group on Mobility of Systems, Users, Data and Computing","SIGOPS ACM Special Interest Group on Operating Systems"]},"container-title":["Proceedings of the 21st Annual International Conference on Mobile Systems, Applications and Services"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581791.3596830","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3581791.3596830","content-type":"text\/html","content-version":"vor","intended-application":"syndication"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:30Z","timestamp":1750178190000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581791.3596830"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,18]]},"references-count":72,"alternative-id":["10.1145\/3581791.3596830","10.1145\/3581791"],"URL":"https:\/\/doi.org\/10.1145\/3581791.3596830","relation":{},"subject":[],"published":{"date-parts":[[2023,6,18]]},"assertion":[{"value":"2023-06-18","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}