{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T17:57:28Z","timestamp":1767117448807,"version":"3.41.0"},"reference-count":47,"publisher":"Association for Computing Machinery (ACM)","issue":"2","license":[{"start":{"date-parts":[[2024,6,20]],"date-time":"2024-06-20T00:00:00Z","timestamp":1718841600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM J. Comput. Sustain. Soc."],"published-print":{"date-parts":[[2024,6,30]]},"abstract":"<jats:p>Early detection and treatment of Social Anxiety Disorder (SAD) is crucial. However, current diagnostic methods have several drawbacks, including being time consuming for clinical interviews, susceptible to emotional bias for self-reports, and inconclusive for physiological measures. Our research focuses on a digital approach using acoustic and linguistic features extracted from participants\u2019 \u201cspeech\u201d for diagnosing SAD. Our methodology involves identifying correlations between extracted features and SAD severity, selecting the effective features, and comparing classical machine learning and deep learning methods for predicting SAD. Our results demonstrate that both acoustic and linguistic features outperform deep learning approaches when considered individually. Logistic Regression proves effective for acoustic features, whereas Random Forest excels with linguistic features, achieving the highest accuracy of 85.71%. Our findings pave the way for non-intrusive SAD diagnosing that can be used conveniently anywhere, facilitating early detection.<\/jats:p>","DOI":"10.1145\/3657245","type":"journal-article","created":{"date-parts":[[2024,4,12]],"date-time":"2024-04-12T12:22:08Z","timestamp":1712924528000},"page":"1-19","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Unveiling Social Anxiety: Analyzing Acoustic and Linguistic Traits in Impromptu Speech within a Controlled Study"],"prefix":"10.1145","volume":"2","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1675-7270","authenticated-orcid":false,"given":"Nilesh Kumar","family":"Sahu","sequence":"first","affiliation":[{"name":"Indian Institute of Science Education and Research Bhopal, Bhopal, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-5768-330X","authenticated-orcid":false,"given":"Manjeet","family":"Yadav","sequence":"additional","affiliation":[{"name":"Indian Institute of Science Education and Research Bhopal, Bhopal, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1245-2974","authenticated-orcid":false,"given":"Haroon R","family":"Lone","sequence":"additional","affiliation":[{"name":"Indian Institute of Science Education and Research Bhopal, Bhopal, India"}]}],"member":"320","published-online":{"date-parts":[[2024,6,20]]},"reference":[{"key":"e_1_3_2_2_2","unstructured":"Psychology Tools. 2022. Social Phobia Inventory (SPIN). Retrieved August 25 2023 from https:\/\/psychology-tools.com\/test\/spin"},{"key":"e_1_3_2_3_2","article-title":"Spectro-temporal acoustical markers differentiate speech from song across cultures","author":"Albouy Philippe","year":"2023","unstructured":"Philippe Albouy, Samuel A. Mehr, Roxane S. Hoyer, Jeremie Ginzburg, and Robert J. Zatorre. 2023. Spectro-temporal acoustical markers differentiate speech from song across cultures. bioRxiv. Published Online, January 29, 2023.","journal-title":"bioRxiv."},{"key":"e_1_3_2_4_2","article-title":"The vocal signature of social anxiety: Exploration using hypothesis-testing and machine-learning approaches","author":"Alon-Ronen Or","year":"2022","unstructured":"Or Alon-Ronen, Yosi Shrem, Yossi Keshet, and Eva Gilboa-Schechtman. 2022. The vocal signature of social anxiety: Exploration using hypothesis-testing and machine-learning approaches. arXiv preprint arXiv:2207.08534 (2022).","journal-title":"arXiv preprint arXiv:2207.08534"},{"key":"e_1_3_2_5_2","unstructured":"American Psychological Association. 2024. Anxiety. Retrieved September 3 2023 from https:\/\/www.apa.org\/topics\/anxiety\/"},{"issue":"2","key":"e_1_3_2_6_2","doi-asserted-by":"crossref","first-page":"240","DOI":"10.1038\/s41562-022-01489-2","article-title":"Deep learning reveals what vocal bursts express in different cultures","volume":"7","author":"Brooks Jeffrey A.","year":"2023","unstructured":"Jeffrey A. Brooks, Panagiotis Tzirakis, Alice Baird, Lauren Kim, Michael Opara, Xia Fang, Dacher Keltner, Maria Monroy, Rebecca Corona, Jacob Metrick, and Alan S. Cowen. 2023. Deep learning reveals what vocal bursts express in different cultures. Nature Human Behaviour 7, 2 (2023), 240\u2013250.","journal-title":"Nature Human Behaviour"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jml.2022.104376"},{"issue":"4","key":"e_1_3_2_8_2","doi-asserted-by":"crossref","first-page":"379","DOI":"10.1192\/bjp.176.4.379","article-title":"Psychometric properties of the Social Phobia Inventory (SPIN): New self-rating scale","volume":"176","author":"Connor Kathryn M.","year":"2000","unstructured":"Kathryn M. Connor, Jonathan R. T. Davidson, L. Erik Churchill, Andrew Sherwood, Richard H. Weisler, and Edna Foa. 2000. Psychometric properties of the Social Phobia Inventory (SPIN): New self-rating scale. British Journal of Psychiatry 176, 4 (2000), 379\u2013386.","journal-title":"British Journal of Psychiatry"},{"issue":"8","key":"e_1_3_2_9_2","doi-asserted-by":"crossref","first-page":"e28918","DOI":"10.2196\/28918","article-title":"Automated screening for social anxiety, generalized anxiety, and depression from objective smartphone-collected data: Cross-sectional study","volume":"23","author":"Matteo Daniel Di","year":"2021","unstructured":"Daniel Di Matteo, Kathryn Fotinos, Sachinthya Lokuge, Geneva Mason, Tia Sternat, Martin A. Katzman, and Jonathan Rose. 2021. Automated screening for social anxiety, generalized anxiety, and depression from objective smartphone-collected data: Cross-sectional study. Journal of Medical Internet Research 23, 8 (2021), e28918.","journal-title":"Journal of Medical Internet Research"},{"issue":"1","key":"e_1_3_2_10_2","doi-asserted-by":"crossref","first-page":"e22723","DOI":"10.2196\/22723","article-title":"Smartphone-detected ambient speech and self-reported measures of anxiety and depression: Exploratory observational study","volume":"5","author":"Matteo Daniel Di","year":"2021","unstructured":"Daniel Di Matteo, Wendy Wang, Kathryn Fotinos, Sachinthya Lokuge, Julia Yu, Tia Sternat, Martin A. Katzman, and Jonathan Rose. 2021. Smartphone-detected ambient speech and self-reported measures of anxiety and depression: Exploratory observational study. JMIR Formative Research 5, 1 (2021), e22723.","journal-title":"JMIR Formative Research"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1145\/3266302.3268997"},{"key":"e_1_3_2_12_2","doi-asserted-by":"crossref","first-page":"372","DOI":"10.1109\/ICOCO53166.2021.9673562","volume-title":"Proceedings of the 2021 IEEE International Conference on Computing (ICOCO \u201921)","author":"Elfaki Ayman","year":"2021","unstructured":"Ayman Elfaki, Ani Liza Asnawi, Ahmad Zamani Jusoh, Ahmad Fadzil Ismail, Siti Noorjannah Ibrahim, Nor Fadhillah Mohamed Azmin, and Nik Nur Wahidah Binti Nik Hashim. 2021. Using the short-time Fourier transform and resnet to diagnose depression from speech data. In Proceedings of the 2021 IEEE International Conference on Computing (ICOCO \u201921). IEEE, 372\u2013376."},{"issue":"6","key":"e_1_3_2_13_2","doi-asserted-by":"crossref","first-page":"651","DOI":"10.1521\/jscp.2013.32.6.651","article-title":"Acoustic properties of dominance and request utterances in social anxiety","volume":"32","author":"Galili Lior","year":"2013","unstructured":"Lior Galili, Ofer Amir, and Eva Gilboa-Schechtman. 2013. Acoustic properties of dominance and request utterances in social anxiety. Journal of Social and Clinical Psychology 32, 6 (2013), 651\u2013673.","journal-title":"Journal of Social and Clinical Psychology"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1080\/10253890.2016.1265939"},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.5121\/sipij.2013.4408"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jad.2018.05.001"},{"key":"e_1_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.5121\/ijdkp.2015.5201"},{"issue":"9","key":"e_1_3_2_18_2","doi-asserted-by":"crossref","first-page":"e0239133","DOI":"10.1371\/journal.pone.0239133","article-title":"Social anxiety in young people: A prevalence study in seven countries","volume":"15","author":"Jefferies Philip","year":"2020","unstructured":"Philip Jefferies and Michael Ungar. 2020. Social anxiety in young people: A prevalence study in seven countries. PLoS One 15, 9 (2020), e0239133.","journal-title":"PLoS One"},{"issue":"2","key":"e_1_3_2_19_2","doi-asserted-by":"crossref","first-page":"251","DOI":"10.1016\/j.scijus.2023.01.004","article-title":"What does method validation look like for forensic voice comparison by a human expert?","volume":"63","author":"Kirchh\u00fcbel Christin","year":"2023","unstructured":"Christin Kirchh\u00fcbel, Georgina Brown, and Paul Foulkes. 2023. What does method validation look like for forensic voice comparison by a human expert? Science & Justice 63, 2 (2023), 251\u2013257.","journal-title":"Science & Justice"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1145\/3544549.3585778"},{"key":"e_1_3_2_21_2","first-page":"1052","volume-title":"Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM \u201917)","author":"Long Hailiang","year":"2017","unstructured":"Hailiang Long, Zhenghao Guo, Xia Wu, Bin Hu, Zhenyu Liu, and Hanshu Cai. 2017. Detecting depression in speech: Comparison and combination between different speech types. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM \u201917). IEEE, 1052\u20131058."},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1002\/lio2.354"},{"key":"e_1_3_2_23_2","first-page":"35","volume-title":"Proceedings of the 6th International Workshop on Audio\/Visual Emotion Challenge","author":"Ma Xingchen","year":"2016","unstructured":"Xingchen Ma, Hongyu Yang, Qiang Chen, Di Huang, and Yunhong Wang. 2016. DepAudioNet: An efficient deep model for audio based depression classification. In Proceedings of the 6th International Workshop on Audio\/Visual Emotion Challenge. 35\u201342."},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.25080\/Majora-7b98e3ed-003"},{"issue":"6","key":"e_1_3_2_25_2","doi-asserted-by":"crossref","first-page":"2294","DOI":"10.1109\/JBHI.2019.2913590","article-title":"Giving voice to vulnerable children: Machine learning analysis of speech detects anxiety and depression in early childhood","volume":"23","author":"McGinnis Ellen W.","year":"2019","unstructured":"Ellen W. McGinnis, Steven P. Anderau, Jessica Hruschak, Reed D. Gurchiek, Nestor L. Lopez-Duran, Kate Fitzgerald, Katherine L. Rosenblum, Maria Muzik, and Ryan S. McGinnis. 2019. Giving voice to vulnerable children: Machine learning analysis of speech detects anxiety and depression in early childhood. IEEE Journal of Biomedical and Health Informatics 23, 6 (2019), 2294\u20132301.","journal-title":"IEEE Journal of Biomedical and Health Informatics"},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2019.2913590"},{"key":"e_1_3_2_27_2","doi-asserted-by":"crossref","first-page":"670","DOI":"10.1109\/SLT48900.2021.9383491","volume-title":"Proceedings of the 2021 IEEE Spoken Language Technology Workshop (SLT \u201921)","author":"Meghanani Amit","year":"2021","unstructured":"Amit Meghanani, C. S. Anoop, and A. G. Ramakrishnan. 2021. An exploration of log-mel spectrogram and MFCC features for Alzheimer\u2019s dementia recognition from spontaneous speech. In Proceedings of the 2021 IEEE Spoken Language Technology Workshop (SLT \u201921). IEEE, 670\u2013677."},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/2676690.2676694"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jneuroling.2006.04.001"},{"key":"e_1_3_2_30_2","unstructured":"Nilesh K. Sahu Manjeet Yadav and Haroon R. Lone. 2024. Social Anxiety Disorder\u2014Participants Audio Data Set. Retrieved April 19 2024 from https:\/\/osf.io\/b8uyg\/"},{"key":"e_1_3_2_31_2","first-page":"625","article-title":"Voice traces of anxiety: Acoustic parameters affected by anxiety disorder","author":"\u00d6zseven Turgut","year":"2018","unstructured":"Turgut \u00d6zseven, Muharrem D\u00fc\u011fenci, Ali Doruk, and Hilal I. Kahraman. 2018. Voice traces of anxiety: Acoustic parameters affected by anxiety disorder. Archives of Acoustics 43, 4 (2018), 625\u2013636.","journal-title":"Archives of Acoustics"},{"key":"e_1_3_2_32_2","volume-title":"Analysis of CNN-Based Speech Recognition System using Raw Speech as Input","author":"Palaz Dimitri","year":"2015","unstructured":"Dimitri Palaz, Ronan Collobert, et\u00a0al. 2015. Analysis of CNN-Based Speech Recognition System using Raw Speech as Input. Technical Report. Idiap Research Institute."},{"key":"e_1_3_2_33_2","article-title":"Linguistic Inquiry and Word Count: LIWC 2001","author":"Pennebaker James W.","year":"2001","unstructured":"James W. Pennebaker, Martha E. Francis, and Roger J. Booth. 2001. Linguistic Inquiry and Word Count: LIWC 2001. Lawrence Erlbaum Associates, Mahwah, NJ.","journal-title":"Lawrence Erlbaum Associates"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1145\/3411823"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1145\/3214284"},{"key":"e_1_3_2_36_2","doi-asserted-by":"crossref","first-page":"1211","DOI":"10.21437\/SpeechProsody.2016-249","article-title":"Do social anxiety individuals hesitate more? The prosodic profile of hesitation disfluencies in social anxiety disorder individuals","author":"Silber-Varod Vered","year":"2016","unstructured":"Vered Silber-Varod, Hamutal Kreiner, Ronen Lovett, Yossi Levi-Belz, and Noam Amir. 2016. Do social anxiety individuals hesitate more? The prosodic profile of hesitation disfluencies in social anxiety disorder individuals. Speech Prosody 2016 (2016), 1211\u20131215.","journal-title":"Speech Prosody 2016"},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2020.04.003"},{"key":"e_1_3_2_38_2","volume-title":"Audio Engineering Society Convention 126","author":"Stein Michael","year":"2009","unstructured":"Michael Stein, Benjamin M. Schubert, Matthias Gruhne, Gabriel Gatzsche, and Markus Mehnert. 2009. Evaluation and comparison of audio chroma feature extraction methods. In Audio Engineering Society Convention 126. Audio Engineering Society."},{"key":"e_1_3_2_39_2","doi-asserted-by":"publisher","DOI":"10.1016\/S0140-6736(08)60488-2"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1145\/3544793.3563406"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.2196\/39998"},{"issue":"7","key":"e_1_3_2_42_2","doi-asserted-by":"crossref","first-page":"e36828","DOI":"10.2196\/36828","article-title":"Acoustic and linguistic features of impromptu speech and their association with anxiety: Validation study","volume":"9","author":"Teferra Bazen Gashaw","year":"2022","unstructured":"Bazen Gashaw Teferra, Sophie Borwein, Danielle D. DeSouza, William Simpson, Ludovic Rheault, and Jonathan Rose. 2022. Acoustic and linguistic features of impromptu speech and their association with anxiety: Validation study. JMIR Mental Health 9, 7 (2022), e36828.","journal-title":"JMIR Mental Health"},{"issue":"1","key":"e_1_3_2_43_2","doi-asserted-by":"crossref","first-page":"e44325","DOI":"10.2196\/44325","article-title":"Predicting generalized anxiety disorder from impromptu speech transcripts using context-aware transformer-based neural networks: Model evaluation study","volume":"10","author":"Teferra Bazen Gashaw","year":"2023","unstructured":"Bazen Gashaw Teferra and Jonathan Rose. 2023. Predicting generalized anxiety disorder from impromptu speech transcripts using context-aware transformer-based neural networks: Model evaluation study. JMIR Mental Health 10, 1 (2023), e44325.","journal-title":"JMIR Mental Health"},{"key":"e_1_3_2_44_2","article-title":"Personalized state anxiety detection: An empirical study with linguistic biomarkers and a machine learning pipeline","author":"Wang Zhiyuan","year":"2023","unstructured":"Zhiyuan Wang, Mingyue Tang, Maria A. Larrazabal, Emma R. Toner, Mark Rucker, Congyu Wu, Bethany A. Teachman, Mehdi Boukhechba, and Laura E. Barnes. 2023. Personalized state anxiety detection: An empirical study with linguistic biomarkers and a machine learning pipeline. arXiv preprint arXiv:2304.09928 (2023).","journal-title":"arXiv preprint arXiv:2304.09928"},{"key":"e_1_3_2_45_2","first-page":"1","volume-title":"Proceedings of the 2023 International Conference on Emerging Smart Computing and Informatics (ESCI \u201923)","author":"Watve Shreyasi","year":"2023","unstructured":"Shreyasi Watve, Mangal Patil, and Arundhati Shinde. 2023. Review of features and classification for spoken Indian language recognition using deep learning and machine learning techniques. In Proceedings of the 2023 International Conference on Emerging Smart Computing and Informatics (ESCI \u201923). IEEE, 1\u20136."},{"key":"e_1_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10862-015-9495-9"},{"key":"e_1_3_2_47_2","unstructured":"WHO. 2022. Mental Disorders. Retrieved September 3 2023 from https:\/\/www.who.int\/news-room\/fact-sheets\/detail\/mental-disorders"},{"key":"e_1_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-017-5539-3"}],"container-title":["ACM Journal on Computing and Sustainable Societies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3657245","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3657245","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:17:38Z","timestamp":1750295858000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3657245"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,20]]},"references-count":47,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,6,30]]}},"alternative-id":["10.1145\/3657245"],"URL":"https:\/\/doi.org\/10.1145\/3657245","relation":{},"ISSN":["2834-5533"],"issn-type":[{"type":"electronic","value":"2834-5533"}],"subject":[],"published":{"date-parts":[[2024,6,20]]},"assertion":[{"value":"2023-09-26","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-03-31","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-06-20","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}