{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,16]],"date-time":"2026-03-16T18:49:17Z","timestamp":1773686957991,"version":"3.50.1"},"reference-count":58,"publisher":"Association for Computing Machinery (ACM)","issue":"1","funder":[{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["JP21K11973"],"award-info":[{"award-number":["JP21K11973"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["JP24K02988"],"award-info":[{"award-number":["JP24K02988"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["JP22H00550"],"award-info":[{"award-number":["JP22H00550"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2026,3,16]]},"abstract":"<jats:p>Modern hearables, such as wireless earbuds with transparency mode, are designed to enhance user awareness by relaying ambient sounds. However, this functionality introduces a new attack surface. We present UltrasonicWhisper+, a novel attack that exploits microphone nonlinearity to inject inaudible ultrasound into hearables, resulting in the demodulation of phantom audible sounds delivered directly to the user. Unlike prior ultrasound-based attacks that target voice assistants, our method deceives users themselves by simulating either internal hearable audio or spatial environmental sounds. We evaluate five commercial hearables and find that demodulated sound quality varies by device, with mean opinion scores (MOS) ranging from 1.26 to 3.27 and short-time objective intelligibility (STOI) ranging from 0.44 to 0.75. Behavioral studies show that participants followed an average of 28.0% of false instructions even after being warned about the attack. Moreover, spatialized ultrasonic sounds achieved 65.5% localization accuracy (\u00b145\u00b0), and a false acceptance rate of 28.4% when perceived as ambient sound. These findings demonstrate that users can be reliably deceived via inaudible audio signals, raising serious concerns for safety-critical applications. Our results call for a reexamination of hearable device security and highlight the need for robust countermeasures.<\/jats:p>","DOI":"10.1145\/3789679","type":"journal-article","created":{"date-parts":[[2026,3,16]],"date-time":"2026-03-16T17:51:14Z","timestamp":1773683474000},"page":"1-33","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["UltrasonicWhisper+: Ultrasonic Attacks Generate Phantom Sounds in Your Hearable"],"prefix":"10.1145","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6854-4448","authenticated-orcid":false,"given":"Hiroki","family":"Watanabe","sequence":"first","affiliation":[{"name":"Future University Hakodate, Hakodate, Hokkaido, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2260-3788","authenticated-orcid":false,"given":"Tsutomu","family":"Terada","sequence":"additional","affiliation":[{"name":"Kobe University, Kobe, Hyogo, Japan"}]}],"member":"320","published-online":{"date-parts":[[2026,3,16]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3341163.3347747"},{"key":"e_1_2_1_2_1","volume-title":"Retrieved","year":"2024","unstructured":"Apple. 2024. Listen with spatial audio for AirPods and Beats. Retrieved February, 2024 from https:\/\/support.apple.com\/en-us\/102469"},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/s42486-021-00074-2"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376304"},{"key":"e_1_2_1_5_1","volume-title":"Real-Time Neural Voice Camouflage. (Dec","author":"Chiquier Mia","year":"2021","unstructured":"Mia Chiquier, Chengzhi Mao, and Carl Vondrick. 2021. Real-Time Neural Voice Camouflage. (Dec. 2021). arXiv:2112.07076 [cs.SD]"},{"key":"e_1_2_1_6_1","first-page":"2","article-title":"PPGface: Like What You are Watching? Earphones can \u201cFeel","volume":"6","author":"Choi Seokmin","year":"2022","unstructured":"Seokmin Choi, Yang Gao, Yincheng Jin, Se Jun Kim, Jiyang Li, Wenyao Xu, and Zhanpeng Jin. 2022. PPGface: Like What You are Watching? Earphones can \u201cFeel\u201d Your Facial Expressions. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6, 2 (July 2022), 1\u201332.","journal-title":"Your Facial Expressions. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."},{"key":"e_1_2_1_7_1","first-page":"5","article-title":"Hearables: Here come the: Technology tucked inside your ears will augment your daily life","volume":"56","author":"Crum Poppy","year":"2019","unstructured":"Poppy Crum. 2019. Hearables: Here come the: Technology tucked inside your ears will augment your daily life. IEEE Spectrum 56, 5 (May 2019), 38\u201343.","journal-title":"IEEE Spectrum"},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3495243.3558763"},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3613904.3642095"},{"key":"e_1_2_1_10_1","first-page":"1","article-title":"Voice In Ear: Spoofing-Resistant and Passphrase-Independent Body Sound Authentication","volume":"5","author":"Gao Yang","year":"2021","unstructured":"Yang Gao, Yincheng Jin, Jagmohan Chauhan, Seokmin Choi, Jiyang Li, and Zhanpeng Jin. 2021. Voice In Ear: Spoofing-Resistant and Passphrase-Independent Body Sound Authentication. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1 (March 2021), 1\u201325.","journal-title":"Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351239"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.1903520"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2023.24457"},{"key":"e_1_2_1_14_1","volume-title":"Retrieved","author":"IDC.","year":"2023","unstructured":"IDC. 2023. Wearable Devices Market Insights. Retrieved March 11, 2024 from https:\/\/www.idc.com\/promo\/wearablevendor"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4880856"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TETC.2019.2953041"},{"key":"e_1_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606782"},{"key":"e_1_2_1_18_1","first-page":"2","article-title":"EarCommand: \u201cHearing","volume":"6","author":"Jin Yincheng","year":"2022","unstructured":"Yincheng Jin, Yang Gao, Xuhai Xu, Seokmin Choi, Jiyang Li, Feng Liu, Zhengxiong Li, and Zhanpeng Jin. 2022. EarCommand: \u201cHearing\u201d Your Silent Speech Commands In Ear. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6, 2 (July 2022), 1\u201328.","journal-title":"Your Silent Speech Commands In Ear. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3607822.3614514"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz470"},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581791.3596837"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3498361.3538921"},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/2556288.2557168"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.21437\/interspeech.2021-299"},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4906160"},{"key":"e_1_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4960481"},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/GCCE.2016.7800460"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1101\/2020.03.31.011650arXiv:https:\/\/www.biorxiv.org\/content\/early\/2020\/03\/31\/2020.03.31.011650.full.pdf"},{"key":"e_1_2_1_29_1","volume-title":"Retrieved","author":"U.S. Department of Housing and Urban Development. [n. d.].","year":"2025","unstructured":"U.S. Department of Housing and Urban Development. [n. d.]. The Noise Guidebook. Retrieved October, 2025 from https:\/\/www.huduser.gov\/portal\/portal\/sites\/default\/files\/pdf\/The-Noise-Guidebook.pdf"},{"key":"e_1_2_1_30_1","first-page":"726","article-title":"The Use of Airborne Ultrasonics for Generating Audible Sound Beams","volume":"47","author":"Pompei Frank Joseph","year":"1999","unstructured":"Frank Joseph Pompei. 1999. The Use of Airborne Ultrasonics for Generating Audible Sound Beams. Journal of the Audio Engineering Society 47, 9 (1999), 726\u2013731.","journal-title":"Journal of the Audio Engineering Society"},{"key":"e_1_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1121\/2.0000622"},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550314"},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3081333.3081366"},{"key":"e_1_2_1_34_1","volume-title":"Pyroomacoustics: A Python Package for Audio Room Simulation and Array Processing Algorithms. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 351\u2013355","author":"Scheibler Robin","year":"2018","unstructured":"Robin Scheibler, Eric Bezzam, and Ivan Dokmani\u0107. 2018. Pyroomacoustics: A Python Package for Audio Room Simulation and Array Processing Algorithms. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 351\u2013355."},{"key":"e_1_2_1_35_1","volume-title":"So Real. Retrieved","year":"2024","unstructured":"Sony. 2024. Sony 360 Reality Audio I So Immersive. So Real. Retrieved February, 2024 from https:\/\/electronics.sony.com\/360-reality-audio"},{"key":"e_1_2_1_36_1","volume-title":"International Conference on Digital Audio Effects (DAFx-15)","author":"Spagnol Simone","year":"2015","unstructured":"Simone Spagnol and Federico Avanzini. 2015. Frequency Estimation of the First Pinna Notch in Head-Related Transfer Functions with a Linear Anthropometric Model. In International Conference on Digital Audio Effects (DAFx-15)."},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550281"},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2019.2917582"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2010.5495701"},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606779"},{"key":"e_1_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3613904.3642057"},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3478085"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534592"},{"key":"e_1_2_1_44_1","volume-title":"31st USENIX Security Symposium (USENIX Security 22)","author":"Wang Kai","year":"2022","unstructured":"Kai Wang, Richard Mitev, Chen Yan, Xiaoyu Ji, Ahmad-Reza Sadeghi, and Wenyuan Xu. 2022. {GhostTouch}: Targeted attacks on touchscreens without physical touch. In 31st USENIX Security Symposium (USENIX Security 22). 1543\u20131559."},{"key":"e_1_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534606"},{"key":"e_1_2_1_46_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3448098","article-title":"EarDynamic: An Ear Canal Deformation Based Continuous User Authentication Using In-Ear Wearables","volume":"5","author":"Wang Zi","year":"2021","unstructured":"Zi Wang, Sheng Tan, Linghan Zhang, Yili Ren, Zhi Wang, and Jie Yang. 2021. EarDynamic: An Ear Canal Deformation Based Continuous User Authentication Using In-Ear Wearables. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 1 (March 2021), 1\u201327.","journal-title":"Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."},{"key":"e_1_2_1_47_1","first-page":"1","article-title":"EarSlide: a Secure Ear Wearables Biometric Authentication Based on Acoustic Fingerprint","volume":"8","author":"Wang Zi","year":"2024","unstructured":"Zi Wang, Yilin Wang, and Jie Yang. 2024. EarSlide: a Secure Ear Wearables Biometric Authentication Based on Acoustic Fingerprint. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 8, 1 (March 2024), 1\u201329.","journal-title":"Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3384657.3384792"},{"key":"e_1_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2023.3312713"},{"key":"e_1_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3594738.3611379"},{"key":"e_1_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1250\/ast.35.159"},{"key":"e_1_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.402445"},{"key":"e_1_2_1_53_1","volume-title":"32nd USENIX Security Symposium (USENIX Security 23)","author":"Xia Qi","year":"2023","unstructured":"Qi Xia, Qian Chen, and Shouhuai Xu. 2023. Near-Ultrasound Inaudible Trojan (Nuit): Exploiting Your Speaker to Attack Your Microphone. In 32nd USENIX Security Symposium (USENIX Security 23). USENIX Association, Anaheim, CA, 4589\u20134606. https:\/\/www.usenix.org\/conference\/usenixsecurity23\/presentation\/xia"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.sna.2005.04.037"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3631120"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134052"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3594738.3611365"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3678546"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3789679","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,16]],"date-time":"2026-03-16T17:52:09Z","timestamp":1773683529000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3789679"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,16]]},"references-count":58,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,3,16]]}},"alternative-id":["10.1145\/3789679"],"URL":"https:\/\/doi.org\/10.1145\/3789679","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,16]]},"assertion":[{"value":"2026-03-16","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}