{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T19:24:46Z","timestamp":1776281086764,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":49,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,13]],"date-time":"2024-05-13T00:00:00Z","timestamp":1715558400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100006374","name":"New Energy and Industrial Technology Development Organization","doi-asserted-by":"publisher","award":["JPNP20006"],"award-info":[{"award-number":["JPNP20006"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]},{"name":"The University of Manchester"},{"name":"Artificial Intelligence Research Center, National Institute of Advanced Industrial Science and Technology"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,13]]},"DOI":"10.1145\/3589334.3648137","type":"proceedings-article","created":{"date-parts":[[2024,5,8]],"date-time":"2024-05-08T07:08:13Z","timestamp":1715152093000},"page":"4489-4500","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":118,"title":["MentaLLaMA: Interpretable Mental Health Analysis on Social Media with Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3142-2516","authenticated-orcid":false,"given":"Kailai","family":"Yang","sequence":"first","affiliation":[{"name":"The University of Manchester, Manchester, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0843-1916","authenticated-orcid":false,"given":"Tianlin","family":"Zhang","sequence":"additional","affiliation":[{"name":"The University of Manchester, Manchester, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-4725-8520","authenticated-orcid":false,"given":"Ziyan","family":"Kuang","sequence":"additional","affiliation":[{"name":"Jiangxi Normal University, Nanchang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9588-7454","authenticated-orcid":false,"given":"Qianqian","family":"Xie","sequence":"additional","affiliation":[{"name":"The University of Manchester, Manchester, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3501-3907","authenticated-orcid":false,"given":"Jimin","family":"Huang","sequence":"additional","affiliation":[{"name":"Wuhan University, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4097-9191","authenticated-orcid":false,"given":"Sophia","family":"Ananiadou","sequence":"additional","affiliation":[{"name":"The University of Manchester, Manchester, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2024,5,13]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/MIS.2023.3254179"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702613.2732733"},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W17-1612"},{"key":"e_1_3_2_2_4_1","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. Advances in neural information processing systems Vol. 33 (2020) 1877--1901."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1023\/A:1021316409277"},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/W15-1204"},{"key":"e_1_3_2_2_7_1","first-page":"16344","article-title":"Flashattention: Fast and memory-efficient exact attention with io-awareness","volume":"35","author":"Dao Tri","year":"2022","unstructured":"Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher R\u00e9. 2022. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, Vol. 35 (2022), 16344--16359.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19--1423"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"crossref","unstructured":"Sara Evans-Lacko Sergio Aguilar-Gaxiola A Al-Hamzawi et al. 2018. Socio-economic variations in the mental health treatment gap for people with anxiety mood and substance use disorders: results from the WHO World Mental Health (WMH) surveys. Psychological medicine Vol. 48 9 (2018) 1560--1571.","DOI":"10.1017\/S0033291717003336"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11831-022-09863-z"},{"key":"e_1_3_2_2_11_1","volume-title":"Proceedings of the Thirteenth Language Resources and Evaluation Conference. European Language Resources Association","author":"Garg Muskan","year":"2022","unstructured":"Muskan Garg, Chandni Saxena, Sriparna Saha, Veena Krishnan, Ruchi Joshi, and Vijay Mago. 2022. CAMS: An Annotated Corpus for Causal Analysis of Mental Health Issues in Social Media Posts. In Proceedings of the Thirteenth Language Resources and Evaluation Conference. European Language Resources Association, Marseille, France, 6387--6396. https:\/\/aclanthology.org\/2022.lrec-1.686"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.757"},{"key":"e_1_3_2_2_13_1","volume-title":"ChatGPT Perpetuates Gender Bias in Machine Translation and Ignores Non-Gendered Pronouns: Findings across Bengali and Five other Low-Resource Languages. arXiv preprint arXiv:2305.10510","author":"Ghosh Sourojit","year":"2023","unstructured":"Sourojit Ghosh and Aylin Caliskan. 2023. ChatGPT Perpetuates Gender Bias in Machine Translation and Ignores Non-Gendered Pronouns: Findings across Bengali and Five other Low-Resource Languages. arXiv preprint arXiv:2305.10510 (2023)."},{"key":"e_1_3_2_2_14_1","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics. International Committee on Computational Linguistics, Gyeongju, Republic of Korea, 94--104","author":"Han Sooji","year":"2022","unstructured":"Sooji Han, Rui Mao, and Erik Cambria. 2022. Hierarchical Attention Network for Explainable Depression Detection on Twitter Aided by Metaphor Concept Mappings. In Proceedings of the 29th International Conference on Computational Linguistics. International Committee on Computational Linguistics, Gyeongju, Republic of Korea, 94--104. https:\/\/aclanthology.org\/2022.coling-1.9"},{"key":"e_1_3_2_2_15_1","volume-title":"MedAlpaca--An Open-Source Collection of Medical Conversational AI Models and Training Data. arXiv preprint arXiv:2304.08247","author":"Han Tianyu","year":"2023","unstructured":"Tianyu Han, Lisa C Adams, Jens-Michalis Papaioannou, Paul Grundmann, Tom Oberhauser, Alexander L\u00f6ser, Daniel Truhn, and Keno K Bressem. 2023. MedAlpaca--An Open-Source Collection of Medical Conversational AI Models and Training Data. arXiv preprint arXiv:2304.08247 (2023)."},{"key":"e_1_3_2_2_16_1","first-page":"3774","article-title":"Do models of mental health based on social media data generalize?. In Findings of the association for computational linguistics","volume":"2020","author":"Harrigian Keith","year":"2020","unstructured":"Keith Harrigian, Carlos Aguirre, and Mark Dredze. 2020. Do models of mental health based on social media data generalize?. In Findings of the association for computational linguistics: EMNLP 2020. 3774--3788.","journal-title":"EMNLP"},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-emnlp.297"},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-021-06208-y"},{"key":"e_1_3_2_2_19_1","volume-title":"Proceedings of the Thirteenth Language Resources and Evaluation Conference. European Language Resources Association","author":"Ji Shaoxiong","year":"2022","unstructured":"Shaoxiong Ji, Tianlin Zhang, Luna Ansari, Jie Fu, Prayag Tiwari, and Erik Cambria. 2022b. MentalBERT: Publicly Available Pretrained Language Models for Mental Healthcare. In Proceedings of the Thirteenth Language Resources and Evaluation Conference. European Language Resources Association, Marseille, France, 7184--7190. https:\/\/aclanthology.org\/2022.lrec-1.778"},{"key":"e_1_3_2_2_20_1","volume-title":"Domain-specific Continued Pretraining of Language Models for Capturing Long Context in Mental Health. arXiv preprint arXiv:2304.10447","author":"Ji Shaoxiong","year":"2023","unstructured":"Shaoxiong Ji, Tianlin Zhang, Kailai Yang, Sophia Ananiadou, Erik Cambria, and J\u00f6rg Tiedemann. 2023. Domain-specific Continued Pretraining of Language Models for Capturing Long Context in Mental Health. arXiv preprint arXiv:2304.10447 (2023)."},{"key":"e_1_3_2_2_21_1","volume-title":"Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa.","author":"Kojima Takeshi","year":"2022","unstructured":"Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. Advances in neural information processing systems, Vol. 35 (2022), 22199--22213."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_2_2_23_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_2_24_1","volume-title":"Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101","author":"Loshchilov Ilya","year":"2017","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)."},{"key":"e_1_3_2_2_25_1","volume-title":"Sad: A stress annotated dataset for recognizing everyday stressors in sms-like conversational systems. In Extended abstracts of the 2021 CHI conference on human factors in computing systems. 1--7.","author":"Mauriello Matthew Louis","year":"2021","unstructured":"Matthew Louis Mauriello, Thierry Lincoln, Grace Hon, Dorien Simon, Dan Jurafsky, and Pablo Paredes. 2021. Sad: A stress annotated dataset for recognizing everyday stressors in sms-like conversational systems. In Extended abstracts of the 2021 CHI conference on human factors in computing systems. 1--7."},{"key":"e_1_3_2_2_26_1","volume-title":"Harlan M Krumholz, Jure Leskovec, Eric J Topol, and Pranav Rajpurkar.","author":"Moor Michael","year":"2023","unstructured":"Michael Moor, Oishi Banerjee, Zahra Shakeri Hossein Abad, Harlan M Krumholz, Jure Leskovec, Eric J Topol, and Pranav Rajpurkar. 2023. Foundation models for generalist medical artificial intelligence. Nature, Vol. 616, 7956 (2023), 259--265."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.578"},{"key":"e_1_3_2_2_28_1","volume-title":"Ethics and privacy in social media research for mental health. Current psychiatry reports","author":"Nicholas Jennifer","year":"2020","unstructured":"Jennifer Nicholas, Sandersan Onie, and Mark E Larsen. 2020. Ethics and privacy in social media research for mental health. Current psychiatry reports, Vol. 22 (2020), 1--7."},{"key":"e_1_3_2_2_29_1","unstructured":"OpenAI. 2023. GPT-4 Technical Report. ArXiv Vol. abs\/2303.08774 (2023). https:\/\/api.semanticscholar.org\/CorpusID:257532815"},{"key":"e_1_3_2_2_30_1","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume":"35","author":"Ouyang Long","year":"2022","unstructured":"Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, Vol. 35 (2022), 27730--27744.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_31_1","volume-title":"Proceedings of the 2018 EMNLP Workshop SMM4H: The 3rd Social Media Mining for Health Applications Workshop & Shared Task. 9--12","author":"Pirina Inna","unstructured":"Inna Pirina and cC aug ri cC \u00f6ltekin. 2018. Identifying depression on reddit: The effect of training data. In Proceedings of the 2018 EMNLP Workshop SMM4H: The 3rd Social Media Mining for Health Applications Workshop & Shared Task. 9--12."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_2_2_33_1","volume-title":"MULTIWD: Multiple Wellness Dimensions in Social Media Posts.","author":"Muskan Garg MSVPJ","year":"2023","unstructured":"MSVPJ SATHVIK and Muskan Garg. 2023. MULTIWD: Multiple Wellness Dimensions in Social Media Posts. (2023)."},{"key":"e_1_3_2_2_34_1","first-page":"3008","article-title":"Learning to summarize with human feedback","volume":"33","author":"Stiennon Nisan","year":"2020","unstructured":"Nisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul F Christiano. 2020. Learning to summarize with human feedback. Advances in Neural Information Processing Systems, Vol. 33 (2020), 3008--3021.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_35_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-10997-4_25"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-6213"},{"key":"e_1_3_2_2_38_1","first-page":"605","article-title":"Generating (factual?) narrative summaries of rcts: Experiments with neural multi-document summarization","volume":"2021","author":"Wallace Byron C","year":"2021","unstructured":"Byron C Wallace, Sayantan Saha, Frank Soboczenski, and Iain J Marshall. 2021. Generating (factual?) narrative summaries of rcts: Experiments with neural multi-document summarization. AMIA Summits on Translational Science Proceedings, Vol. 2021 (2021), 605.","journal-title":"AMIA Summits on Translational Science Proceedings"},{"key":"e_1_3_2_2_39_1","volume-title":"International Conference on Machine Learning. PMLR, 22964--22984","author":"Wang Thomas","year":"2022","unstructured":"Thomas Wang, Adam Roberts, Daniel Hesslow, Teven Le Scao, Hyung Won Chung, Iz Beltagy, Julien Launay, and Colin Raffel. 2022. What language model architecture and pretraining objective works best for zero-shot generalization?. In International Conference on Machine Learning. PMLR, 22964--22984."},{"key":"e_1_3_2_2_40_1","volume-title":"Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations, ICLR 2023","author":"Wang Xuezhi","year":"2023","unstructured":"Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc V. Le, Ed H. Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. 2023 b. Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1--5, 2023. OpenReview.net. https:\/\/openreview.net\/pdf?id=1PL1NIMMrw"},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"e_1_3_2_2_42_1","unstructured":"Jason Wei Yi Tay Rishi Bommasani Colin Raffel Barret Zoph Sebastian Borgeaud Dani Yogatama Maarten Bosma Denny Zhou Donald Metzler et al. 2022. Emergent Abilities of Large Language Models. Transactions on Machine Learning Research (2022)."},{"key":"e_1_3_2_2_43_1","volume-title":"PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark for Finance. arXiv preprint arXiv:2306.05443","author":"Xie Qianqian","year":"2023","unstructured":"Qianqian Xie, Weiguang Han, Xiao Zhang, Yanzhao Lai, Min Peng, Alejandro Lopez-Lira, and Jimin Huang. 2023. PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark for Finance. arXiv preprint arXiv:2306.05443 (2023)."},{"key":"e_1_3_2_2_44_1","volume-title":"Leveraging Large Language Models for Mental Health Prediction via Online Text Data. arXiv preprint arXiv:2307.14385","author":"Xu Xuhai","year":"2023","unstructured":"Xuhai Xu, Bingshen Yao, Yuanzhe Dong, Hong Yu, James Hendler, Anind K Dey, and Dakuo Wang. 2023. Leveraging Large Language Models for Mental Health Prediction via Online Text Data. arXiv preprint arXiv:2307.14385 (2023)."},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.370"},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.102961"},{"key":"e_1_3_2_2_47_1","first-page":"27263","article-title":"Bartscore: Evaluating generated text as text generation","volume":"34","author":"Yuan Weizhe","year":"2021","unstructured":"Weizhe Yuan, Graham Neubig, and Pengfei Liu. 2021. Bartscore: Evaluating generated text as text generation. Advances in Neural Information Processing Systems, Vol. 34 (2021), 27263--27277.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2023.103417"},{"key":"e_1_3_2_2_49_1","volume-title":"Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022","author":"Zhang Zhiling","year":"2022","unstructured":"Zhiling Zhang, Siyuan Chen, Mengyue Wu, and Kenny Q. Zhu. 2022. Psychiatric Scale Guided Risky Post Screening for Early Detection of Depression. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23--29 July 2022, Luc De Raedt (Ed.). 5220--5226."}],"event":{"name":"WWW '24: The ACM Web Conference 2024","location":"Singapore Singapore","acronym":"WWW '24","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM Web Conference 2024"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3589334.3648137","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3589334.3648137","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:26:48Z","timestamp":1755822408000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3589334.3648137"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,13]]},"references-count":49,"alternative-id":["10.1145\/3589334.3648137","10.1145\/3589334"],"URL":"https:\/\/doi.org\/10.1145\/3589334.3648137","relation":{},"subject":[],"published":{"date-parts":[[2024,5,13]]},"assertion":[{"value":"2024-05-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}