{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T08:01:45Z","timestamp":1773734505590,"version":"3.50.1"},"reference-count":59,"publisher":"Association for Computing Machinery (ACM)","issue":"2","license":[{"start":{"date-parts":[[2020,6,15]],"date-time":"2020-06-15T00:00:00Z","timestamp":1592179200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key R&D Program of China","doi-asserted-by":"crossref","award":["2018YFC1900700"],"award-info":[{"award-number":["2018YFC1900700"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100012659","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61772338"],"award-info":[{"award-number":["61772338"]}],"id":[{"id":"10.13039\/501100012659","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2020,6,15]]},"abstract":"<jats:p>Recent years have witnessed the surge of biometric-based user authentication for mobile devices due to its promising security and convenience. As a natural and widely-existed behavior, human speaking has been exploited for user authentication. Existing voice-based user authentication explores the unique characteristics from either the voiceprint or mouth movements, which is vulnerable to replay attacks and mimic attacks. During speaking, the vocal tract, including the static shape and dynamic movements, also exhibits the individual uniqueness, and they are hardly eavesdropped and imitated by adversaries. Hence, our work aims to employ the individual uniqueness of vocal tract to realize user authentication on mobile devices. Moreover, most voice-based user authentications are passphrase-dependent, which significantly degrade the user experience. Thus, such user authentications are pressed to be implemented in a passphrase-independent manner while being able to resist various attacks. In this paper, we propose a user authentication system, VocalLock, which senses the whole vocal tract during speaking to identify different individuals in a passphrase-independent manner on smartphones leveraging acoustic signals. VocalLock first utilizes FMCW on acoustic signals to characterize both the static shape and dynamic movements of the vocal tract during speaking, and then constructs a passphrase-independent user authentication model based on the unique characteristics of vocal tract through GMM-UBM. The proposed VocalLock can resist various spoofing attacks, while achieving a satisfactory user experience. Extensive experiments in real environments demonstrate VocalLock can accurately authenticate user identity in a passphrase-independent manner and successfully resist various attacks.<\/jats:p>","DOI":"10.1145\/3397320","type":"journal-article","created":{"date-parts":[[2020,6,15]],"date-time":"2020-06-15T22:30:37Z","timestamp":1592260237000},"page":"1-24","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":33,"title":["VocalLock"],"prefix":"10.1145","volume":"4","author":[{"given":"Li","family":"Lu","sequence":"first","affiliation":[{"name":"Shanghai Jiao Tong University, Department of Computer Science and Engineering, Shanghai, China"}]},{"given":"Jiadi","family":"Yu","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Department of Computer Science and Engineering, Shanghai, China"}]},{"given":"Yingying","family":"Chen","sequence":"additional","affiliation":[{"name":"Rutgers University, WINLAB and Department of Electrical and Computer Engineering, New Brunswick, NJ, USA"}]},{"given":"Yan","family":"Wang","sequence":"additional","affiliation":[{"name":"Temple University, Department of Computer and Information Sciences, Philadelphia, PA, USA"}]}],"member":"320","published-online":{"date-parts":[[2020,6,15]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/PROC.1977.10770"},{"key":"e_1_2_1_2_1","unstructured":"Amazon. 2019. Echo & Alexa - Amazon Device. [Online]. Available: https:\/\/www.amazon.com. (2019).  Amazon. 2019. Echo & Alexa - Amazon Device. [Online]. Available: https:\/\/www.amazon.com. (2019)."},{"key":"e_1_2_1_3_1","unstructured":"Apple. 2019. iPhone XS - FaceID - Apple. [Online]. Available: https:\/\/www.apple.com\/iphone-xs\/face-id\/. (2019).  Apple. 2019. iPhone XS - FaceID - Apple. [Online]. Available: https:\/\/www.apple.com\/iphone-xs\/face-id\/. (2019)."},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCA.2010.2041656"},{"key":"e_1_2_1_5_1","unstructured":"C. BYU. 2020. Word frequency: based on 450 million word coca corpus. [Online]. Available: https:\/\/www.wordfrequency.info. (2020).  C. BYU. 2020. Word frequency: based on 450 million word coca corpus. [Online]. Available: https:\/\/www.wordfrequency.info. (2020)."},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.628714"},{"key":"e_1_2_1_7_1","first-page":"1","article-title":"An Analysis of Power Consumption in a Smartphone. In Proc. USENIX ATC. Boston","volume":"21","author":"Carroll Aaron","year":"2010","unstructured":"Aaron Carroll and Gernot Heiser . 2010 . An Analysis of Power Consumption in a Smartphone. In Proc. USENIX ATC. Boston , MA, USA , 21 : 1 -- 21 :14. Aaron Carroll and Gernot Heiser. 2010. An Analysis of Power Consumption in a Smartphone. In Proc. USENIX ATC. Boston, MA, USA, 21:1--21:14.","journal-title":"MA, USA"},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3314390"},{"key":"e_1_2_1_9_1","volume-title":"Proc. IEEE ICDCS. 183--195","author":"Chen S.","unstructured":"S. Chen , K. Ren , S. Piao , C. Wang , Q. Wang , J. Weng , L. Su , and A. Mohaisen . 2017. You Can Hear But You Cannot Steal: Defending Against Voice Impersonation Attacks on Smartphones . In Proc. IEEE ICDCS. 183--195 . S. Chen, K. Ren, S. Piao, C. Wang, Q. Wang, J. Weng, L. Su, and A. Mohaisen. 2017. You Can Hear But You Cannot Steal: Defending Against Voice Impersonation Attacks on Smartphones. In Proc. IEEE ICDCS. 183--195."},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2064307"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.2517-6161.1977.tb01600.x"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/PROC.1985.13345"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/89.279278"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/79.317924"},{"key":"e_1_2_1_15_1","volume-title":"Proc. AISTATS'12","author":"Glorot Xavier","year":"2012","unstructured":"Xavier Glorot , Antoine Bordes , Yoshua Bengio , Xavier Glorot , Antoine Bordes , and Yoshua Bengio . 2012 . Deep Sparse Rectifier Neural Networks . In Proc. AISTATS'12 . La Palma, Canary Islands, 315--323. Xavier Glorot, Antoine Bordes, Yoshua Bengio, Xavier Glorot, Antoine Bordes, and Yoshua Bengio. 2012. Deep Sparse Rectifier Neural Networks. In Proc. AISTATS'12. La Palma, Canary Islands, 315--323."},{"key":"e_1_2_1_16_1","unstructured":"Google. 2019. Google Home - Smart Speaker & Home Assistant. [Online]. Available: https:\/\/store.google.com\/us\/product\/google_home. (2019).  Google. 2019. Google Home - Smart Speaker & Home Assistant. [Online]. Available: https:\/\/store.google.com\/us\/product\/google_home. (2019)."},{"key":"e_1_2_1_17_1","unstructured":"Google. 2019. Google Smart Lock. [Online]. Available: https:\/\/get.google.com\/smartlock\/. (2019).  Google. 2019. Google Smart Lock. [Online]. Available: https:\/\/get.google.com\/smartlock\/. (2019)."},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2014.05.021"},{"key":"e_1_2_1_19_1","volume-title":"Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167","author":"Ioffe Sergey","year":"2015","unstructured":"Sergey Ioffe and Christian Szegedy . 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167 ( 2015 ). Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167 (2015)."},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1002\/sec.1499"},{"key":"e_1_2_1_22_1","volume-title":"International journal of human-computer studies 65, 1","author":"Keith Mark","year":"2007","unstructured":"Mark Keith , Benjamin Shao , and Paul John Steinbart . 2007. The usability of passphrases for authentication: An empirical field study . International journal of human-computer studies 65, 1 ( 2007 ), 17--28. Mark Keith, Benjamin Shao, and Paul John Steinbart. 2007. The usability of passphrases for authentication: An empirical field study. International journal of human-computer studies 65, 1 (2007), 17--28."},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/PROC.1967.5962"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.21236\/ADA613971"},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978397"},{"key":"e_1_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2019.2891733"},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2018.8486283"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351255"},{"key":"e_1_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2019.8737591"},{"key":"e_1_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/2973750.2973755"},{"key":"e_1_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3210240.3210325"},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2011.5947436"},{"key":"e_1_2_1_33_1","volume-title":"Proc","author":"Mukhopadhyay Dibya","unstructured":"Dibya Mukhopadhyay , Maliheh Shirvanian , and Nitesh Saxena . 2015. All Your Voices are Belong to Us: Stealing Voices to Fool Humans and Machines . In Proc . ESORICS. Springer , Vienna, Austria , 599--621. Dibya Mukhopadhyay, Maliheh Shirvanian, and Nitesh Saxena. 2015. All Your Voices are Belong to Us: Stealing Voices to Fool Humans and Machines. In Proc. ESORICS. Springer, Vienna, Austria, 599--621."},{"key":"e_1_2_1_34_1","volume-title":"Proc. ISCA INTERSPEECH","author":"Nagrani A.","unstructured":"A. Nagrani , J. S. Chung , and A. Zisserman . 2017. VoxCeleb: a large-scale speaker identification dataset . In Proc. ISCA INTERSPEECH . Stockholm, Sweden, 2616--2620. A. Nagrani, J. S. Chung, and A. Zisserman. 2017. VoxCeleb: a large-scale speaker identification dataset. In Proc. ISCA INTERSPEECH. Stockholm, Sweden, 2616--2620."},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3214278"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351258"},{"key":"e_1_2_1_37_1","volume-title":"Acousticcardiogram: Monitoring Heartbeats using Acoustic Signals on Smart Devices. In Proc","author":"Qian K.","year":"2018","unstructured":"K. Qian , C. Wu , F. Xiao , Y. Zheng , Y. Zhang , Z. Yang , and Y. Liu . 2018 . Acousticcardiogram: Monitoring Heartbeats using Acoustic Signals on Smart Devices. In Proc . IEEE INFOCOM. Honolulu, HI , USA , 1574--1582. K. Qian, C. Wu, F. Xiao, Y. Zheng, Y. Zhang, Z. Yang, and Y. Liu. 2018. Acousticcardiogram: Monitoring Heartbeats using Acoustic Signals on Smart Devices. In Proc. IEEE INFOCOM. Honolulu, HI, USA, 1574--1582."},{"key":"e_1_2_1_38_1","volume-title":"Proc. ISCA EUROSPEECH","author":"Reynolds Douglas A.","year":"1997","unstructured":"Douglas A. Reynolds . 1997 . Comparison of Background Normalization Methods for Text-Independent Speaker Verification . In Proc. ISCA EUROSPEECH . Rhodes, Greece, 963--966. Douglas A. Reynolds. 1997. Comparison of Background Normalization Methods for Text-Independent Speaker Verification. In Proc. ISCA EUROSPEECH. Rhodes, Greece, 963--966."},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1006\/dspr.1999.0361"},{"key":"e_1_2_1_40_1","unstructured":"Samsung. 2017. Iris recognition on Galaxy S8. [Online]. Available: https:\/\/www.samsung.com\/au\/iris\/. (2017).  Samsung. 2017. Iris recognition on Galaxy S8. [Online]. Available: https:\/\/www.samsung.com\/au\/iris\/. (2017)."},{"key":"e_1_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2010.5495503"},{"key":"e_1_2_1_43_1","volume-title":"Proc. ISMIR","author":"Sigurdsson Sigurdur","year":"2006","unstructured":"Sigurdur Sigurdsson , Kaare Brandt Petersen , and Tue Lehn-Schi\u00f8ler . 2006 . Mel Frequency Cepstral Coefficients: An Evaluation of Robustness of MP3 Encoded Music . In Proc. ISMIR . Victoria, Canada, 286--289. Sigurdur Sigurdsson, Kaare Brandt Petersen, and Tue Lehn-Schi\u00f8ler. 2006. Mel Frequency Cepstral Coefficients: An Evaluation of Robustness of MP3 Encoded Music. In Proc. ISMIR. Victoria, Canada, 286--289."},{"key":"e_1_2_1_44_1","volume-title":"Radar handbook","author":"Skolnik Merrill Ivan","unstructured":"Merrill Ivan Skolnik . 1970. Radar handbook . McGraw-Hill, Incorporated , New York, NY, USA . Merrill Ivan Skolnik. 1970. Radar handbook. McGraw-Hill, Incorporated, New York, NY, USA."},{"key":"e_1_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2017.8057099"},{"key":"e_1_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3191768"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/2493190.2493231"},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3161188"},{"key":"e_1_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICMLC.2011.6016982"},{"key":"e_1_2_1_50_1","volume-title":"Voiceprint: The New Wechat Password. [Online]. Available: https:\/\/blog.wechat.com\/2015\/05\/21\/voiceprint-the-new-wechat-password\/.","year":"2015","unstructured":"Wechat. 2015 . Voiceprint: The New Wechat Password. [Online]. Available: https:\/\/blog.wechat.com\/2015\/05\/21\/voiceprint-the-new-wechat-password\/. (2015). Wechat. 2015. Voiceprint: The New Wechat Password. [Online]. Available: https:\/\/blog.wechat.com\/2015\/05\/21\/voiceprint-the-new-wechat-password\/. (2015)."},{"key":"e_1_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2014.10.005"},{"key":"e_1_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2017.8057022"},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3307334.3326074"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/3319535.3354248"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2004.81"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3081333.3081356"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126474"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3133962"},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978296"},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3243734.3243777"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3397320","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3397320","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T23:12:47Z","timestamp":1750201967000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3397320"}},"subtitle":["Sensing Vocal Tract for Passphrase-Independent User Authentication Leveraging Acoustic Signals on Smartphones"],"short-title":[],"issued":{"date-parts":[[2020,6,15]]},"references-count":59,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2020,6,15]]}},"alternative-id":["10.1145\/3397320"],"URL":"https:\/\/doi.org\/10.1145\/3397320","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,6,15]]},"assertion":[{"value":"2020-06-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}