{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T21:35:34Z","timestamp":1770845734077,"version":"3.50.1"},"reference-count":15,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T00:00:00Z","timestamp":1765756800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T00:00:00Z","timestamp":1765756800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,15]]},"DOI":"10.1109\/bibm66473.2025.11356556","type":"proceedings-article","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T21:19:40Z","timestamp":1769721580000},"page":"7638-7644","source":"Crossref","is-referenced-by-count":0,"title":["Multi-Strategy Collaborative Training of Large Language Models for Depression Text Generation and Detection"],"prefix":"10.1109","author":[{"given":"Zhengnan","family":"Qiao","sequence":"first","affiliation":[{"name":"College of Computer Science, Inner Mongolia University,Hohhot,China"}]},{"given":"Xianguo","family":"Zhang","sequence":"additional","affiliation":[{"name":"College of Computer Science, Inner Mongolia University,Hohhot,China"}]}],"member":"263","reference":[{"key":"ref1","volume-title":"Depression","year":"2023"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.21437\/interspeech.2019-2036"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.3390\/bdcc8090112"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1322"},{"key":"ref5","article-title":"LoRA: Low-Rank Adaptation of Large Language Models","author":"Hu","year":"2021","journal-title":"arXiv preprint"},{"key":"ref6","article-title":"Finetuned Language Models are Zero-Shot Learners","volume-title":"presented at the Int. Conf. Learn. Represent. (ICLR)","author":"Wei","year":"2022"},{"key":"ref7","article-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach","volume-title":"presented at the Int. Conf. Learn. Represent. (ICLR)","author":"Liu","year":"2020"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1410"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.2196\/22635"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/w18-5903"},{"key":"ref11","first-page":"3123","article-title":"The Distress Analysis Interview Corpus of human and computer interviews","author":"Gratch","year":"2014","journal-title":"Proc. LREC"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.65109\/mxiv3169"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref14","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2022.107503"}],"event":{"name":"2025 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)","location":"Wuhan, China","start":{"date-parts":[[2025,12,15]]},"end":{"date-parts":[[2025,12,18]]}},"container-title":["2025 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11355913\/11355975\/11356556.pdf?arnumber=11356556","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T20:51:32Z","timestamp":1770843092000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11356556\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,15]]},"references-count":15,"URL":"https:\/\/doi.org\/10.1109\/bibm66473.2025.11356556","relation":{},"subject":[],"published":{"date-parts":[[2025,12,15]]}}}