{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T06:26:40Z","timestamp":1769927200240,"version":"3.49.0"},"reference-count":37,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2021,12,27]],"date-time":"2021-12-27T00:00:00Z","timestamp":1640563200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100006952","name":"Louisiana Board of Regents","doi-asserted-by":"publisher","award":["LEQSF(2020-23)-RD-A-11"],"award-info":[{"award-number":["LEQSF(2020-23)-RD-A-11"]}],"id":[{"id":"10.13039\/100006952","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2021,12,27]]},"abstract":"<jats:p>The ability to identify pedestrians unobtrusively is essential for smart buildings to provide customized environments, energy saving, health monitoring and security-enhanced services. In this paper, we present an unobtrusive pedestrian identification system by passively listening to people's walking sounds. The proposed acoustic system can be easily integrated with the widely deployed voice assistant devices while providing the context awareness ability. This work focuses on two major tasks. Firstly, we address the challenge of recognizing footstep sounds in complex indoor scenarios by exploiting deep learning and the advanced stereo recording technology that is available on most voice assistant devices. We develop a Convolutional Neural Network-based algorithm and the footstep sound-oriented signal processing schemes to identify users by their footstep sounds accurately. Secondly, we design a \"live\" footstep detection approach to defend against replay attacks. By deriving the novel inter-footstep and intra-footstep characteristics, we distinguish live footstep sounds from the machine speaker's replay sounds based on their spatial variances. The system is evaluated under normal scenarios, traditional replay attacks and the advanced replays, which are designed to forge footstep sounds both acoustically and spatially. Extensive experiments show that our system identifies people with up to 94.9% accuracy in one footstep and shields 100% traditional replay attacks and up to 99% advanced replay attacks.<\/jats:p>","DOI":"10.1145\/3494963","type":"journal-article","created":{"date-parts":[[2021,12,30]],"date-time":"2021-12-30T17:40:33Z","timestamp":1640886033000},"page":"1-19","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["Unobtrusive Pedestrian Identification by Leveraging Footstep Sounds with Replay Resistance"],"prefix":"10.1145","volume":"5","author":[{"given":"Long","family":"Huang","sequence":"first","affiliation":[{"name":"Louisiana State University, Baton Rouge, Louisiana, USA"}]},{"given":"Chen","family":"Wang","sequence":"additional","affiliation":[{"name":"Louisiana State University, Baton Rouge, Louisiana, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,12,30]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4830853"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2007.902460"},{"key":"e_1_2_1_3_1","first-page":"018","article-title":"System for detecting sound generation position and method for detecting sound generation position","volume":"16","author":"Asano Taishi","year":"2020","unstructured":"Taishi Asano , Shunsuke Saito , Naofumi Kitano , and Kenji Tachibana . 2020 . System for detecting sound generation position and method for detecting sound generation position . US Patent App. 16\/586 , 018 . Taishi Asano, Shunsuke Saito, Naofumi Kitano, and Kenji Tachibana. 2020. System for detecting sound generation position and method for detecting sound generation position. US Patent App. 16\/586,018.","journal-title":"US Patent App."},{"key":"e_1_2_1_4_1","unstructured":"Wendy Bumgardner. 2019. What Are the Average Steps Per Minute for Different Exercises? https:\/\/www.verywellfit.com\/pedometer-step-equivalents-for-exercises-and-activities-3435742.  Wendy Bumgardner. 2019. What Are the Average Steps Per Minute for Different Exercises? https:\/\/www.verywellfit.com\/pedometer-step-equivalents-for-exercises-and-activities-3435742."},{"key":"e_1_2_1_5_1","volume-title":"25th {USENIX} Security Symposium ({USENIX} Security 16). 513--530.","author":"Carlini Nicholas","unstructured":"Nicholas Carlini , Pratyush Mishra , Tavish Vaidya , Yuankai Zhang , Micah Sherr , Clay Shields , David Wagner , and Wenchao Zhou . 2016. Hidden voice commands . In 25th {USENIX} Security Symposium ({USENIX} Security 16). 513--530. Nicholas Carlini, Pratyush Mishra, Tavish Vaidya, Yuankai Zhang, Micah Sherr, Clay Shields, David Wagner, and Wenchao Zhou. 2016. Hidden voice commands. In 25th {USENIX} Security Symposium ({USENIX} Security 16). 513--530."},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/SPW.2018.00009"},{"key":"e_1_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS.2017.133"},{"key":"e_1_2_1_8_1","unstructured":"Phillip L De Leon Michael Pucher and Junichi Yamagishi. 2010. Evaluation of the vulnerability of speaker verification to synthetic speech. (2010).  Phillip L De Leon Michael Pucher and Junichi Yamagishi. 2010. Evaluation of the vulnerability of speaker verification to synthetic speech. (2010)."},{"key":"e_1_2_1_9_1","volume-title":"Accelerometer-based gait analysis, a survey. Nor Informasjonssikkerhetskonferanse NISK","author":"Derawi Mohammad Omar","year":"2010","unstructured":"Mohammad Omar Derawi . 2010. Accelerometer-based gait analysis, a survey. Nor Informasjonssikkerhetskonferanse NISK ( 2010 ). Mohammad Omar Derawi. 2010. Accelerometer-based gait analysis, a survey. Nor Informasjonssikkerhetskonferanse NISK (2010)."},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2668024.2668027"},{"key":"e_1_2_1_11_1","unstructured":"Dave Gershgorn. 2017. Alexa doesn't have the attention span to secretly eavesdrop on your conversations. https:\/\/qz.com\/1121880\/the-technical-reason-why-alexa-cant-listen-into-your-private-conversations\/.  Dave Gershgorn. 2017. Alexa doesn't have the attention span to secretly eavesdrop on your conversations. https:\/\/qz.com\/1121880\/the-technical-reason-why-alexa-cant-listen-into-your-private-conversations\/."},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2015.05.002"},{"key":"e_1_2_1_13_1","unstructured":"Google Nest Help. 2020. More about data security and privacy on devices that work with Assistant. https:\/\/support.google.com\/googlenest\/answer\/7072285?hl=en.  Google Nest Help. 2020. More about data security and privacy on devices that work with Assistant. https:\/\/support.google.com\/googlenest\/answer\/7072285?hl=en."},{"key":"e_1_2_1_14_1","volume-title":"Fabio Di Troia, and Mark Stamp","author":"Huang Jasper","year":"2018","unstructured":"Jasper Huang , Fabio Di Troia, and Mark Stamp . 2018 . A Comparison of Machine Learning Classifiers for Acoustic Gait Analysis. In Proceedings of the International Conference on Security and Management (SAM). The Steering Committee of The World Congress in Computer Science, Computer Engineering and Applied Computing (WorldComp) , 160--166. Jasper Huang, Fabio Di Troia, and Mark Stamp. 2018. A Comparison of Machine Learning Classifiers for Acoustic Gait Analysis. In Proceedings of the International Conference on Security and Management (SAM). The Steering Committee of The World Congress in Computer Science, Computer Engineering and Applied Computing (WorldComp), 160--166."},{"key":"e_1_2_1_15_1","unstructured":"Tricia Hussung. 2017. Cell Phone Addiction: The Statistics of Gadget Dependency. https:\/\/online.king.edu\/news\/cell-phone-addiction\/  Tricia Hussung. 2017. Cell Phone Addiction: The Statistics of Gadget Dependency. https:\/\/online.king.edu\/news\/cell-phone-addiction\/"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.21437\/Eurospeech.1999-283"},{"key":"e_1_2_1_17_1","unstructured":"miniDSP. 2020. UMA-8 USB mic array V2.0. https:\/\/www.minidsp.com\/products\/usb-audio-interface\/uma-8-microphone-array.  miniDSP. 2020. UMA-8 USB mic array V2.0. https:\/\/www.minidsp.com\/products\/usb-audio-interface\/uma-8-microphone-array."},{"key":"e_1_2_1_18_1","volume-title":"Combining evidence from residual phase and MFCC features for speaker recognition","author":"Murty K Sri Rama","year":"2005","unstructured":"K Sri Rama Murty and Bayya Yegnanarayana . 2005. Combining evidence from residual phase and MFCC features for speaker recognition . IEEE signal processing letters 13, 1 ( 2005 ), 52--55. K Sri Rama Murty and Bayya Yegnanarayana. 2005. Combining evidence from residual phase and MFCC features for speaker recognition. IEEE signal processing letters 13, 1 (2005), 52--55."},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2006.886018"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/633292.633453"},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2699343.2699364"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2014.20"},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2009.4960661"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISSNIP.2009.5416778"},{"key":"e_1_2_1_25_1","unstructured":"Olivia Solon. 2020. Facial recognition bill would ban use by federal law enforcement. https:\/\/www.nbcnews.com\/tech\/security\/2-democratic-senators-propose-ban-use-facial-recognition-federal-law-n1232128.  Olivia Solon. 2020. Facial recognition bill would ban use by federal law enforcement. https:\/\/www.nbcnews.com\/tech\/security\/2-democratic-senators-propose-ban-use-facial-recognition-federal-law-n1232128."},{"key":"e_1_2_1_26_1","volume-title":"An overview of speaker identification: Accuracy and robustness issues","author":"Togneri Roberto","year":"2011","unstructured":"Roberto Togneri and Daniel Pullella . 2011. An overview of speaker identification: Accuracy and robustness issues . IEEE circuits and systems magazine 11, 2 ( 2011 ), 23--61. Roberto Togneri and Daniel Pullella. 2011. An overview of speaker identification: Accuracy and robustness issues. IEEE circuits and systems magazine 11, 2 (2011), 23--61."},{"key":"e_1_2_1_27_1","volume-title":"9th {USENIX} Workshop on Offensive Technologies ({WOOT} 15).","author":"Vaidya Tavish","unstructured":"Tavish Vaidya , Yuankai Zhang , Micah Sherr , and Clay Shields . 2015. Cocaine noodles: exploiting the gap between human and machine speech recognition . In 9th {USENIX} Workshop on Offensive Technologies ({WOOT} 15). Tavish Vaidya, Yuankai Zhang, Micah Sherr, and Clay Shields. 2015. Cocaine noodles: exploiting the gap between human and machine speech recognition. In 9th {USENIX} Workshop on Offensive Technologies ({WOOT} 15)."},{"key":"e_1_2_1_28_1","volume-title":"Julian Fierrez, and Javier Ortega-Garcia.","author":"Vera-Rodriguez Ruben","year":"2012","unstructured":"Ruben Vera-Rodriguez , John SD Mason , Julian Fierrez, and Javier Ortega-Garcia. 2012 . Comparative analysis and fusion of spatiotemporal information for footstep recognition. IEEE transactions on pattern analysis and machine intelligence 35, 4 (2012), 823--834. Ruben Vera-Rodriguez, John SD Mason, Julian Fierrez, and Javier Ortega-Garcia. 2012. Comparative analysis and fusion of spatiotemporal information for footstep recognition. IEEE transactions on pattern analysis and machine intelligence 35, 4 (2012), 823--834."},{"key":"e_1_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.future.2017.07.012"},{"key":"e_1_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2013.6639067"},{"key":"e_1_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3308558.3313591"},{"key":"e_1_2_1_32_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3351273","article-title":"Acousticid: gait-based human identification using acoustic signal","volume":"3","author":"Xu Wei","year":"2019","unstructured":"Wei Xu , ZhiWen Yu , Zhu Wang , Bin Guo , and Qi Han . 2019 . Acousticid: gait-based human identification using acoustic signal . Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 3 , 3 (2019), 1 -- 25 . Wei Xu, ZhiWen Yu, Zhu Wang, Bin Guo, and Qi Han. 2019. Acousticid: gait-based human identification using acoustic signal. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 3, 3 (2019), 1--25.","journal-title":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"},{"key":"e_1_2_1_33_1","volume-title":"MU-ID: Multi-user Identification Through Gaits Using Millimeter Wave Radios. In IEEE INFOCOM 2020-IEEE Conference on Computer Communications. IEEE, 2589--2598","author":"Yang Xin","year":"2020","unstructured":"Xin Yang , Jian Liu , Yingying Chen , Xiaonan Guo , and Yucheng Xie . 2020 . MU-ID: Multi-user Identification Through Gaits Using Millimeter Wave Radios. In IEEE INFOCOM 2020-IEEE Conference on Computer Communications. IEEE, 2589--2598 . Xin Yang, Jian Liu, Yingying Chen, Xiaonan Guo, and Yucheng Xie. 2020. MU-ID: Multi-user Identification Through Gaits Using Millimeter Wave Radios. In IEEE INFOCOM 2020-IEEE Conference on Computer Communications. IEEE, 2589--2598."},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.5555\/2959355.2959359"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134052"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3133962"},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978296"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3494963","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3494963","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:59Z","timestamp":1750191539000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3494963"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12,27]]},"references-count":37,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2021,12,27]]}},"alternative-id":["10.1145\/3494963"],"URL":"https:\/\/doi.org\/10.1145\/3494963","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,12,27]]},"assertion":[{"value":"2021-12-30","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}