{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T11:07:56Z","timestamp":1770462476953,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":29,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,10,18]],"date-time":"2021-10-18T00:00:00Z","timestamp":1634515200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Kunshan Government Research (KGR) Funding in AY 2020\/2021"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,10,18]]},"DOI":"10.1145\/3461615.3491111","type":"proceedings-article","created":{"date-parts":[[2021,12,18]],"date-time":"2021-12-18T04:57:40Z","timestamp":1639803460000},"page":"104-111","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Call For Help Detection In Emergent Situations Using Keyword Spotting And Paralinguistic Analysis"],"prefix":"10.1145","author":[{"given":"Huangrui","family":"Chu","sequence":"first","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, China"}]},{"given":"Yechen","family":"Wang","sequence":"additional","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, China"}]},{"given":"Ran","family":"Ju","sequence":"additional","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, China"}]},{"given":"Yan","family":"Jia","sequence":"additional","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, China"}]},{"given":"Haoxu","family":"Wang","sequence":"additional","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, China"}]},{"given":"Ming","family":"Li","sequence":"additional","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, China"}]},{"given":"Qi","family":"Deng","sequence":"additional","affiliation":[{"name":"Technology Asia and Escalator, KONE Elevators Co., Ltd., China"}]}],"member":"320","published-online":{"date-parts":[[2021,12,17]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"aidatatang_200zh, a free Chinese Mandarin speech corpus.Retrieved","author":"Ltd Beijing DataTang Technology\u00a0Co.","year":"2021","unstructured":"Ltd Beijing DataTang Technology\u00a0Co. 2006. aidatatang_200zh, a free Chinese Mandarin speech corpus.Retrieved Sep 1, 2021 from http:\/\/www.openslr.org\/62\/ Ltd Beijing DataTang Technology\u00a0Co.2006. aidatatang_200zh, a free Chinese Mandarin speech corpus.Retrieved Sep 1, 2021 from http:\/\/www.openslr.org\/62\/"},{"key":"e_1_3_2_1_2_1","unstructured":"Hui Bu Jiayu Du Xingyu Na Bengu Wu and Hao Zheng. 2017. AISHELL-1: An Open-Source Mandarin Speech Corpus and A Speech Recognition Baseline. CoRR abs\/1709.05522(2017). arXiv:1709.05522http:\/\/arxiv.org\/abs\/1709.05522  Hui Bu Jiayu Du Xingyu Na Bengu Wu and Hao Zheng. 2017. AISHELL-1: An Open-Source Mandarin Speech Corpus and A Speech Recognition Baseline. CoRR abs\/1709.05522(2017). arXiv:1709.05522http:\/\/arxiv.org\/abs\/1709.05522"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2013.6639336"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2014.6854370"},{"key":"e_1_3_2_1_5_1","unstructured":"K. Chlasta K. Wo\u0142k and I. Krejtz. 2019. Automated speech-based screening of depression using deep convolutional neural networks. ArXiv abs\/1912.01115(2019).  K. Chlasta K. Wo\u0142k and I. Krejtz. 2019. Automated speech-based screening of depression using deep convolutional neural networks. ArXiv abs\/1912.01115(2019)."},{"key":"#cr-split#-e_1_3_2_1_6_1.1","doi-asserted-by":"crossref","unstructured":"David Crystal. 2019. PARALINGUISTICS. De Gruyter Mouton 265-296. https:\/\/doi.org\/10.1515\/9783111659916-008 10.1515\/9783111659916-008","DOI":"10.1515\/9783111659916-008"},{"key":"#cr-split#-e_1_3_2_1_6_1.2","doi-asserted-by":"crossref","unstructured":"David Crystal. 2019. PARALINGUISTICS. De Gruyter Mouton 265-296. https:\/\/doi.org\/10.1515\/9783111659916-008","DOI":"10.1515\/9783111659916-008"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2064307"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2010.2064307"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/1014052.1014067"},{"key":"e_1_3_2_1_10_1","volume-title":"Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","author":"He Kaiming","year":"2016","unstructured":"Kaiming He , X. Zhang , Shaoqing Ren , and Jian Sun . 2016 . Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 770\u2013778. Kaiming He, X. Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 770\u2013778."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7953152"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2015.10.020"},{"key":"e_1_3_2_1_13_1","volume-title":"Magic Data Technology\u00a0Co.2019. Retrieved","year":"2021","unstructured":"[ 13 ] Ltd. Magic Data Technology\u00a0Co.2019. Retrieved Sep 1, 2021 from http:\/\/www.imagicdatatech.com\/index.php\/home\/dataopensource\/data_info\/id\/101 [13] Ltd. Magic Data Technology\u00a0Co.2019. Retrieved Sep 1, 2021 from http:\/\/www.imagicdatatech.com\/index.php\/home\/dataopensource\/data_info\/id\/101"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Soo\u00a0Jin Park G. Yeung J. Kreiman P. Keating and A. Alwan. 2017. Using Voice Quality Features to Improve Short-Utterance Text-Independent Speaker Verification Systems. In INTERSPEECH.  Soo\u00a0Jin Park G. Yeung J. Kreiman P. Keating and A. Alwan. 2017. Using Voice Quality Features to Improve Short-Utterance Text-Independent Speaker Verification Systems. In INTERSPEECH.","DOI":"10.21437\/Interspeech.2017-157"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"Daniel Povey Gaofeng Cheng Yiming Wang Ke Li Hainan Xu Mahsa Yarmohammadi and Sanjeev Khudanpur. 2018. Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks.. In Interspeech. 3743\u20133747.  Daniel Povey Gaofeng Cheng Yiming Wang Ke Li Hainan Xu Mahsa Yarmohammadi and Sanjeev Khudanpur. 2018. Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks.. In Interspeech. 3743\u20133747.","DOI":"10.21437\/Interspeech.2018-1417"},{"key":"e_1_3_2_1_16_1","unstructured":"Daniel Povey A. Ghoshal Gilles Boulianne L. Burget O. Glembek N. Goel M. Hannemann P. Motl\u00edcek Y. Qian Petr Schwarz J. Silovsk\u00fd G. Stemmer and Karel Vesel\u00fd. 2011. The Kaldi Speech Recognition Toolkit.  Daniel Povey A. Ghoshal Gilles Boulianne L. Burget O. Glembek N. Goel M. Hannemann P. Motl\u00edcek Y. Qian Petr Schwarz J. Silovsk\u00fd G. Stemmer and Karel Vesel\u00fd. 2011. The Kaldi Speech Recognition Toolkit."},{"key":"e_1_3_2_1_17_1","unstructured":"Ltd. Primewords Information Technology\u00a0Co.2018. Primewords Chinese Corpus Set 1. https:\/\/www.primewords.cn.  Ltd. Primewords Information Technology\u00a0Co.2018. Primewords Chinese Corpus Set 1. https:\/\/www.primewords.cn."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"T. Sainath and Carolina Parada. 2015. Convolutional neural networks for small-footprint keyword spotting. In INTERSPEECH.  T. Sainath and Carolina Parada. 2015. Convolutional neural networks for small-footprint keyword spotting. In INTERSPEECH.","DOI":"10.21437\/Interspeech.2015-352"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2012.02.005"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Changhao Shan Junbo Zhang Yujun Wang and L. Xie. 2018. Attention-based End-to-End Models for Small-Footprint Keyword Spotting. In INTERSPEECH.  Changhao Shan Junbo Zhang Yujun Wang and L. Xie. 2018. Attention-based End-to-End Models for Small-Footprint Keyword Spotting. In INTERSPEECH.","DOI":"10.21437\/Interspeech.2018-1777"},{"key":"e_1_3_2_1_21_1","unstructured":"Yao Shi Hui Bu Xin Xu Shaoji Zhang and Ming Li. 2020. AISHELL-3: A Multi-speaker Mandarin TTS Corpus and the Baselines. CoRR abs\/2010.11567(2020). arXiv:2010.11567https:\/\/arxiv.org\/abs\/2010.11567  Yao Shi Hui Bu Xin Xu Shaoji Zhang and Ming Li. 2020. AISHELL-3: A Multi-speaker Mandarin TTS Corpus and the Baselines. CoRR abs\/2010.11567(2020). arXiv:2010.11567https:\/\/arxiv.org\/abs\/2010.11567"},{"key":"e_1_3_2_1_22_1","volume-title":"MUSAN: A Music, Speech, and Noise Corpus. CoRR abs\/1510.08484(2015). arXiv:1510.08484http:\/\/arxiv.org\/abs\/1510.08484","author":"Snyder David","year":"2015","unstructured":"David Snyder , Guoguo Chen , and Daniel Povey . 2015 . MUSAN: A Music, Speech, and Noise Corpus. CoRR abs\/1510.08484(2015). arXiv:1510.08484http:\/\/arxiv.org\/abs\/1510.08484 David Snyder, Guoguo Chen, and Daniel Povey. 2015. MUSAN: A Music, Speech, and Noise Corpus. CoRR abs\/1510.08484(2015). arXiv:1510.08484http:\/\/arxiv.org\/abs\/1510.08484"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"David Snyder D. Garcia-Romero Daniel Povey and S. Khudanpur. 2017. Deep Neural Network Embeddings for Text-Independent Speaker Verification. In INTERSPEECH.  David Snyder D. Garcia-Romero Daniel Povey and S. Khudanpur. 2017. Deep Neural Network Embeddings for Text-Independent Speaker Verification. In INTERSPEECH.","DOI":"10.21437\/Interspeech.2017-620"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2016.7846306"},{"key":"e_1_3_2_1_25_1","unstructured":"SurfingTech.[n.d.]. ST-CMDS-20170001_1 Free ST Chinese Mandarin Corpus. https:\/\/www.openslr.org\/38\/.  SurfingTech.[n.d.]. ST-CMDS-20170001_1 Free ST Chinese Mandarin Corpus. https:\/\/www.openslr.org\/38\/."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"Martin Vondra and R. V\u00edch. 2008. Evaluation of Speech Emotion Classification Based on GMM and Data Fusion. In COST 2102 Conference.  Martin Vondra and R. V\u00edch. 2008. Evaluation of Speech Emotion Classification Based on GMM and Data Fusion. In COST 2102 Conference.","DOI":"10.1007\/978-3-642-03320-9_10"},{"key":"e_1_3_2_1_27_1","unstructured":"Dong Wang and Xuewei Zhang. 2015. THCHS-30 : A Free Chinese Speech Corpus. CoRR abs\/1512.01882(2015). arXiv:1512.01882http:\/\/arxiv.org\/abs\/1512.01882  Dong Wang and Xuewei Zhang. 2015. THCHS-30 : A Free Chinese Speech Corpus. CoRR abs\/1512.01882(2015). arXiv:1512.01882http:\/\/arxiv.org\/abs\/1512.01882"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Peng Yang C. Leung Lei Xie B. Ma and Haizhou Li. 2014. Intrinsic spectral analysis based on temporal context features for query-by-example spoken term detection. In INTERSPEECH.  Peng Yang C. Leung Lei Xie B. Ma and Haizhou Li. 2014. Intrinsic spectral analysis based on temporal context features for query-by-example spoken term detection. In INTERSPEECH.","DOI":"10.21437\/Interspeech.2014-394"}],"event":{"name":"ICMI '21: INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION","location":"Montreal QC Canada","acronym":"ICMI '21","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Companion Publication of the 2021 International Conference on Multimodal Interaction"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3461615.3491111","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3461615.3491111","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:49:04Z","timestamp":1750193344000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3461615.3491111"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,18]]},"references-count":29,"alternative-id":["10.1145\/3461615.3491111","10.1145\/3461615"],"URL":"https:\/\/doi.org\/10.1145\/3461615.3491111","relation":{},"subject":[],"published":{"date-parts":[[2021,10,18]]},"assertion":[{"value":"2021-12-17","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}