{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T12:26:24Z","timestamp":1769775984692,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":20,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556397","type":"print"},{"value":"9789819556403","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5640-3_21","type":"book-chapter","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T21:07:48Z","timestamp":1769720868000},"page":"322-337","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["TCM-R1: Enhancing the\u00a0Traditional Chinese Medicine Capabilities of\u00a0Large Language Models Through Group Relative Policy Optimization"],"prefix":"10.1007","author":[{"given":"Weiwei","family":"He","sequence":"first","affiliation":[]},{"given":"Song","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Li","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shiyi","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Qi","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,30]]},"reference":[{"key":"21_CR1","doi-asserted-by":"crossref","unstructured":"Ren, M., et al.: TCM-SD: a benchmark for probing syndrome differentiation via natural language processing. In: Proceedings of the 21st Chinese National Conference on Computational Linguistics (CCL) (2022)","DOI":"10.1007\/978-3-031-18315-7_16"},{"key":"21_CR2","doi-asserted-by":"crossref","unstructured":"Zhao, Z., et al.: Thinking the importance of patient\u2019s chief complaint in TCM syndrome differentiation. In: Proceedings of the 27th International Conference on Computer Supported Cooperative Work Design (CSCWD) (2024)","DOI":"10.1109\/CSCWD61410.2024.10580801"},{"key":"21_CR3","doi-asserted-by":"crossref","unstructured":"Pampari, A., Raghavan, P., Liang, J., Peng, J.: emrQA: a large corpus for question answering on electronic medical records. In: Proceedings of the Conference on Empirical Methods Natural Language Processing (EMNLP) (2018)","DOI":"10.18653\/v1\/D18-1258"},{"key":"21_CR4","doi-asserted-by":"crossref","unstructured":"Li, D., Hu, B., Chen, Q., Peng, T.: Towards medical machine reading comprehension with structural knowledge and plain text. In: Proceedings of the Conference on Empirical Methods Natural Language Processing (EMNLP) (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.111"},{"key":"21_CR5","doi-asserted-by":"crossref","unstructured":"Jochim, C., Deleris, L.: Named entity recognition in the medical domain with constrained CRF models. In: Proceedings of the 15th Conference on European Chapter of the Association for Computational Linguistics (EACL) (2017)","DOI":"10.18653\/v1\/E17-1079"},{"key":"21_CR6","unstructured":"Gu, Y., Dong, L., Wei, F., Huang, M.: MiniLLM: knowledge distillation of large language models. In: Proceedings of the 12th International Conference on Learning Representations (ICLR) (2024)"},{"key":"21_CR7","doi-asserted-by":"crossref","unstructured":"Hsieh, C.Y., et al.: Distilling step-by-step! Outperforming larger language models with less training data and smaller model sizes. In: Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.507"},{"key":"21_CR8","unstructured":"Lu, P., et al.: T-SciQ: teaching multimodal chain-of-thought reasoning via mixed large language model signals for science question answering. arXiv:2405.10737 (2024)"},{"key":"21_CR9","unstructured":"Zhang, Y., Li, X., Wang, J.: Qibo: a large language model for traditional Chinese medicine. OpenReview preprint, (2024). https:\/\/openreview.net\/forum?id=ftWc6BW77Iy"},{"key":"21_CR10","doi-asserted-by":"crossref","unstructured":"Hua, L., Chen, Q., Lu, H.: Lingdan: enhancing encoding of traditional Chinese medicine knowledge for clinical reasoning tasks with large language models. arXiv:2403.16056 (2024)","DOI":"10.1093\/jamia\/ocae087"},{"key":"21_CR11","doi-asserted-by":"crossref","unstructured":"Wei, J., Li, X., Wang, J.: BianCang: a traditional Chinese medicine large language model. arXiv:2411.11027 (2024)","DOI":"10.1109\/JBHI.2025.3612415"},{"key":"21_CR12","doi-asserted-by":"crossref","unstructured":"Tan, Y., Li, X., Wang, J.: MedChatZH: a tuning LLM for traditional Chinese medicine consultations. Unpublished preprint (2024)","DOI":"10.1016\/j.compbiomed.2024.108290"},{"key":"21_CR13","unstructured":"Gunasekar, S., Zhang, Y., Aneja, J., Mendes, C.C.T.: Textbooks are all you need. arXiv:2306.11644 (2023)"},{"key":"21_CR14","doi-asserted-by":"crossref","unstructured":"Dai, Y., Shao, X., et al.: TCMChat: a generative large language model for traditional Chinese medicine. Unpublished preprint (2024)","DOI":"10.1016\/j.phrs.2024.107530"},{"key":"21_CR15","unstructured":"Chen, Z., Zhong, W., Chen, K., et al.: HuatuoGPT-II, one stage training for medical adaption of LLMs. arXiv:2310.05621 (2023)"},{"key":"21_CR16","unstructured":"Cui, Y., Yang, Z., Yao, X., et al.: Efficient and effective text encoding for Chinese LLaMA and Alpaca. arXiv:2304.08177 (2023)"},{"key":"21_CR17","unstructured":"DeepSeek-AI, D., Guo, D., Yang, et al.: DeepSeek-R1: incentivizing reasoning capability in LLMs via reinforcement learning. arXiv:2501.12948 (2025)"},{"key":"21_CR18","unstructured":"Wei, Z., Yue, W., Wang, X.: ShenNong-TCM: a traditional Chinese medicine large language model (2023). https:\/\/github.com\/michael-wzhu\/ShenNong-TCM-LLM"},{"key":"21_CR19","unstructured":"Yang, A., et al.: Baichuan 2: Open large-scale language models. arXiv:2309.10305 (2023)"},{"key":"21_CR20","doi-asserted-by":"crossref","unstructured":"Yang, S., et al.: Zhongjing: enhancing the Chinese medical capabilities of large language model through expert feedback and real-world multi-turn dialogue. In: Proceedings of the AAAI Conference on Artificial Intelligence (2024)","DOI":"10.1609\/aaai.v38i17.29907"}],"container-title":["Lecture Notes in Computer Science","Web and Big Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5640-3_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T21:07:54Z","timestamp":1769720874000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5640-3_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556397","9789819556403"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5640-3_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"30 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"APWeb-WAIM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asia-Pacific Web (APWeb) and Web-Age Information Management (WAIM) Joint International Conference on Web and Big Data","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenyang","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 August 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"apwebwaim2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/apweb2025.sau.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}