{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:05:36Z","timestamp":1750309536638,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":32,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,12,13]],"date-time":"2024-12-13T00:00:00Z","timestamp":1734048000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,12,13]]},"DOI":"10.1145\/3711542.3711559","type":"proceedings-article","created":{"date-parts":[[2025,4,14]],"date-time":"2025-04-14T04:46:56Z","timestamp":1744606016000},"page":"168-172","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["AnxiBERT - Prompt Engineering for Social Media Anxiety Detection"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-4634-1657","authenticated-orcid":false,"given":"Ashala Lakmini","family":"Senanayake","sequence":"first","affiliation":[{"name":"Faculty of Computing &amp; Technology, University of Kelaniya, Colombo, Sri Lanka"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0778-0874","authenticated-orcid":false,"given":"Sidath","family":"Liyanage","sequence":"additional","affiliation":[{"name":"Faculty of Computing &amp; Technology, University of Kelaniya, Colombo, Sri Lanka"}]}],"member":"320","published-online":{"date-parts":[[2025,4,13]]},"reference":[{"key":"e_1_3_3_2_1_2","volume-title":"State of Emotional Health Disorders ofUndergraduate Students in the United ArabEmirates: A Cross-Sectional Survey,\" in Psychology Research and Behavior Management","author":"Marzouqi A. M. A.","year":"2022","unstructured":"A. M. A. Marzouqi, M. E. Otim, A. Alblooshi, S. A. Marzooqi, M. Talal and F. Wassim, \"State of Emotional Health Disorders ofUndergraduate Students in the United ArabEmirates: A Cross-Sectional Survey,\" in Psychology Research and Behavior Management, 2022."},{"key":"e_1_3_3_2_2_2","volume-title":"Comparison of machine learning for sentiment analysis in detecting anxiety based on social media data,\" arXiv preprint arXiv:2101.06353","author":"Saifullah S.","year":"2021","unstructured":"S. Saifullah, Y. Fauziah and A. S. Aribowo, \"Comparison of machine learning for sentiment analysis in detecting anxiety based on social media data,\" arXiv preprint arXiv:2101.06353, 2021."},{"key":"e_1_3_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.12720\/jait.13.2.167-172"},{"key":"e_1_3_3_2_4_2","volume-title":"Quantifying the natural sentiment strength of polar term senses using semantic gloss information and degree adverbs,\" Journal of Advances in Information Technology Vol","author":"Darwich M.","year":"2020","unstructured":"M. Darwich, S. A. M. Noah, N. Omar, N. A. Osman and I. Said, \"Quantifying the natural sentiment strength of polar term senses using semantic gloss information and degree adverbs,\" Journal of Advances in Information Technology Vol, vol. 11, 2020."},{"key":"e_1_3_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1007\/s41060-018-0096-z"},{"key":"e_1_3_3_2_6_2","volume-title":"Detecting anxiety through reddit,\" Proceedings of the Fourth Workshop on Computational Linguistics and Clinical Psychology\u2014From Linguistic Signal to Clinical Reality","author":"Shen J. H.","year":"2017","unstructured":"J. H. Shen and F. Rudzicz, \"Detecting anxiety through reddit,\" Proceedings of the Fourth Workshop on Computational Linguistics and Clinical Psychology\u2014From Linguistic Signal to Clinical Reality, 2017."},{"key":"e_1_3_3_2_7_2","first-page":"726","volume-title":"Linguistic correlates of social anxiety disorder,\" Cognition & emotion","author":"Hofmann S. G.","year":"2012","unstructured":"S. G. Hofmann, P. M. Moore, C. Gutner and J. W. Weeks, \"Linguistic correlates of social anxiety disorder,\" Cognition & emotion, vol. 26, no. Taylor & Francis, pp. 720\u2013726, 2012."},{"key":"e_1_3_3_2_8_2","volume-title":"Speech and Signal Processing (ICASSP)","author":"Winata G. I.","year":"2018","unstructured":"G. I. Winata, O. P. Kampman and P. Fung, \"Attention-based lstm for psychological stress detection from spoken language using distant supervision,\" in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2018."},{"key":"e_1_3_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2021.08.002"},{"key":"e_1_3_3_2_10_2","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding,\" arXiv preprint arXiv:1810.04805","author":"Devlin J.","year":"2018","unstructured":"J. Devlin, M.-W. Chang, K. Lee and K. Toutanova, \"Bert: Pre-training of deep bidirectional transformers for language understanding,\" arXiv preprint arXiv:1810.04805, 2018."},{"key":"e_1_3_3_2_11_2","volume-title":"Roberta: A robustly optimized bert pretraining approach,\" arXiv preprint arXiv:1907.11692","author":"Liu Y.","year":"2019","unstructured":"Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer and V. Stoyanov, \"Roberta: A robustly optimized bert pretraining approach,\" arXiv preprint arXiv:1907.11692, 2019."},{"key":"e_1_3_3_2_12_2","first-page":"1077","volume-title":"IEEE","author":"Vajre V.","year":"2021","unstructured":"V. Vajre, M. Naylor, U. Kamath and A. Shehu, \"PsychBERT: a mental health language model for social media mental health behavioral analysis,\" in 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), IEEE, 2021, pp. 1077-1082."},{"key":"e_1_3_3_2_13_2","volume-title":"L. Ansari, J. Fu, P. Tiwari and E. Cambria, \"Mentalbert: Publicly available pretrained language models for mental healthcare,\" arXiv preprint arXiv:2110.15621","author":"S.","year":"2021","unstructured":"S. a. Z. T. Ji, L. Ansari, J. Fu, P. Tiwari and E. Cambria, \"Mentalbert: Publicly available pretrained language models for mental healthcare,\" arXiv preprint arXiv:2110.15621, 2021."},{"key":"e_1_3_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.4038\/jdrra.v1i2.20"},{"key":"e_1_3_3_2_15_2","volume-title":"Finetuned language models are zero-shot learners,\" arXiv preprint arXiv:2109.01652","author":"Wei J.","year":"2021","unstructured":"J. Wei, M. Bosma, V. Y. Zhao, K. Guu, A. W. Yu, B. Lester, N. Du, A. M. Dai and Q. V. Le, \"Finetuned language models are zero-shot learners,\" arXiv preprint arXiv:2109.01652, 2021."},{"key":"e_1_3_3_2_16_2","volume-title":"Albert: A lite bert for self-supervised learning of language representations,\" arXiv preprint arXiv:1909.11942","author":"Lan Z.","year":"2019","unstructured":"Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma and R. Soricut, \"Albert: A lite bert for self-supervised learning of language representations,\" arXiv preprint arXiv:1909.11942, 2019."},{"key":"e_1_3_3_2_17_2","volume-title":"Prompt Engineering for Digital Mental Health: A Short Review,\" Frontiers in Digital Health","author":"Priyadarshana Y.","year":"2024","unstructured":"Y. Priyadarshana, A. Senanayake, Z. Liang and I. Piumarta, \"Prompt Engineering for Digital Mental Health: A Short Review,\" Frontiers in Digital Health, vol. 6, 2024."},{"key":"e_1_3_3_2_18_2","first-page":"3921","article-title":"Sun and others, \"Knowledge inheritance for pre-trained language models,\" in Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics","author":"Qin Y.","year":"2022","unstructured":"Y. Qin, Y. Lin, J. Yi, J. Zhang, X. Han, Z. Zhang, Y. Su, Z. Liu, P. Li, M. Sun and others, \"Knowledge inheritance for pre-trained language models,\" in Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022, pp. 3921\u20133937.","journal-title":"Human Language Technologies"},{"key":"e_1_3_3_2_19_2","unstructured":"X. Liu Y. Zheng Z. Du M. Ding Y. Qian Z. Yang and J. Tang \"GPT understands too \" AI Open."},{"key":"e_1_3_3_2_20_2","first-page":"3045","article-title":"The power of scale for parameter-efficient prompt tuning,\" in Proceedings of the 2021 Conference on Empirical Methods","author":"Lester B.","year":"2021","unstructured":"B. Lester, R. Al-Rfou and N. Constant, \"The power of scale for parameter-efficient prompt tuning,\" in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021, pp. 3045\u20133059.","journal-title":"Natural Language Processing"},{"key":"e_1_3_3_2_21_2","first-page":"1901","volume-title":"Askell and others, \"Language models are few-shot learners,\" Advances in neural information processing systems","author":"Brown T.","year":"2020","unstructured":"T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell and others, \"Language models are few-shot learners,\" Advances in neural information processing systems, vol. 33, pp. 1877\u20131901, 2020."},{"key":"e_1_3_3_2_22_2","volume-title":"N. Goyal, E. Hambro, F. Azhar and others, \"Llama: Open and efficient foundation language models,\" {arXiv preprint arXiv:2302.13971","author":"Touvron H.","year":"2023","unstructured":"H. Touvron, T. Lavril, G. Izacard, X. Martinet, M.-A. Lachaux, T. Lacroix, B. Rozi{`e}re, N. Goyal, E. Hambro, F. Azhar and others, \"Llama: Open and efficient foundation language models,\" {arXiv preprint arXiv:2302.13971, 2023."},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3643540","article-title":"Mental-llm: Leveraging large language models for mental health prediction via online text data","volume":"8","author":"Xu X.","year":"2024","unstructured":"X. Xu, B. Yao, Y. Dong, S. Gabriel, H. Yu, J. Hendler, M. Ghassemi, A. K. Dey and D. Wang, \"Mental-llm: Leveraging large language models for mental health prediction via online text data,\" Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, vol. 8, pp. 1-32, 2024.","journal-title":"Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"},{"key":"e_1_3_3_2_24_2","first-page":"24824","volume-title":"Zhou and others, \"Chain-of-thought prompting elicits reasoning in large language models,\" Advances in neural information processing systems","author":"Wei J.","year":"2022","unstructured":"J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, Q. V. Le, D. Zhou and others, \"Chain-of-thought prompting elicits reasoning in large language models,\" Advances in neural information processing systems, vol. 35, pp. 24824-24837, 2022."},{"key":"e_1_3_3_2_25_2","first-page":"22199","volume-title":"a. G. S. S. a. R. M. a. M. Y. a. I. Y. Kojima, \"Large language models are zero-shot reasoners,\" Advances in neural information processing systems","author":"T.","year":"2022","unstructured":"T. a. G. S. S. a. R. M. a. M. Y. a. I. Y. Kojima, \"Large language models are zero-shot reasoners,\" Advances in neural information processing systems, vol. 35, pp. 22199-22213, 2022."},{"key":"e_1_3_3_2_26_2","first-page":"156","article-title":"How does in-context learning help prompt tuning?","volume":"2024","author":"Sun S.","year":"2024","unstructured":"S. a. L. Y. a. I. D. a. Z. C. a. I. M. Sun, \"How does in-context learning help prompt tuning?,\" in Findings of the Association for Computational Linguistics: EACL 2024, 2024, pp. 156-165.","journal-title":"Findings of the Association for Computational Linguistics: EACL"},{"key":"e_1_3_3_2_27_2","volume-title":"Openprompt: An open-source framework for prompt-learning,\" arXiv preprint arXiv:2111.01998","author":"Ding N.","year":"2021","unstructured":"N. Ding, S. Hu, W. Zhao, Y. Chen, Z. Liu, H.-T. Zheng and M. Sun, \"Openprompt: An open-source framework for prompt-learning,\" arXiv preprint arXiv:2111.01998, 2021."},{"key":"e_1_3_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.635"},{"key":"e_1_3_3_2_29_2","first-page":"8","article-title":"Emotion-based modeling of mental disorders on social media","author":"Guo X.","year":"2021","unstructured":"X. Guo, Y. Sun and S. Vosoughi, \"Emotion-based modeling of mental disorders on social media,\" in IEEE\/WIC\/ACM International Conference on Web Intelligence and Intelligent Agent Technology, 2021, pp. 8-16.","journal-title":"IEEE\/WIC\/ACM International Conference on Web Intelligence and Intelligent Agent Technology"},{"key":"e_1_3_3_2_30_2","first-page":"15305","article-title":"DisorBERT: A double domain adaptation model for detecting signs of mental disorders in social media,\" in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1","author":"Aragon M.","year":"2023","unstructured":"M. Aragon, A. P. L. Monroy, L. Gonzalez, D. E. Losada and M. Montes, \"DisorBERT: A double domain adaptation model for detecting signs of mental disorders in social media,\" in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023, pp. 15305-15318.","journal-title":"Long Papers)"},{"key":"e_1_3_3_2_31_2","volume-title":"A baseline for detecting misclassified and out-of-distribution examples in neural networks,\" arXiv preprint arXiv:1610.02136","author":"Hendrycks D.","year":"2016","unstructured":"D. Hendrycks and K. Gimpel, \"A baseline for detecting misclassified and out-of-distribution examples in neural networks,\" arXiv preprint arXiv:1610.02136, 2016."},{"key":"e_1_3_3_2_32_2","volume-title":"Adam: A method for stochastic optimization,\" arXiv preprint arXiv:1412.6980","author":"Kingma D. P.","year":"2014","unstructured":"D. P. Kingma and J. Ba, \"Adam: A method for stochastic optimization,\" arXiv preprint arXiv:1412.6980, 2014."}],"event":{"name":"NLPIR 2024: 2024 8th International Conference on Natural Language Processing and Information Retrieval","acronym":"NLPIR 2024","location":"Okayama Japan"},"container-title":["Proceedings of the 2024 8th International Conference on Natural Language Processing and Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711542.3711559","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3711542.3711559","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:29Z","timestamp":1750295909000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711542.3711559"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,13]]},"references-count":32,"alternative-id":["10.1145\/3711542.3711559","10.1145\/3711542"],"URL":"https:\/\/doi.org\/10.1145\/3711542.3711559","relation":{},"subject":[],"published":{"date-parts":[[2024,12,13]]},"assertion":[{"value":"2025-04-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}