{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,6]],"date-time":"2026-04-06T05:22:41Z","timestamp":1775452961324,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":44,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819756773","type":"print"},{"value":"9789819756780","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-5678-0_3","type":"book-chapter","created":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T16:04:18Z","timestamp":1722528258000},"page":"25-37","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["LegalGPT: Legal Chain of Thought for the Legal Large Language Model Multi-agent Framework"],"prefix":"10.1007","author":[{"given":"Juanming","family":"Shi","sequence":"first","affiliation":[]},{"given":"Qinglang","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Liao","sequence":"additional","affiliation":[]},{"given":"Shenglin","family":"Liang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,1]]},"reference":[{"key":"3_CR1","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3_CR2","doi-asserted-by":"crossref","unstructured":"Black, S., et al.: GPT-NeoX-20B: an open-source autoregressive language model. arXiv preprint arXiv:2204.06745 (2022)","DOI":"10.18653\/v1\/2022.bigscience-1.9"},{"key":"3_CR3","unstructured":"Zhang, S., et al.: OPT: open pre-trained transformer Language models. arXiv preprint arXiv:2205.01068 (2022)"},{"key":"3_CR4","unstructured":"Smith, S., et al.: Using deepspeed and megatron to train megatron-turing NLG 530b, a large-scale generative language model. arXiv preprint arXiv:2201.11990 (2022)"},{"key":"3_CR5","unstructured":"OpenAI: GPT-4 Technical report. arXiv abs\/2303.08774 (2023)"},{"key":"3_CR6","unstructured":"Penedo, G., et al.: The RefinedWeb dataset for Falcon LLM: outperforming curated corpora with web data, and web data only. arXiv preprint arXiv:2306.01116 (2023)"},{"key":"3_CR7","unstructured":"Anil, R., et al.: PaLM 2 technical report. arXiv preprint arXiv:2305.10403 (2023)"},{"key":"3_CR8","unstructured":"Araci, D.: FinBERT: financial sentiment analysis with pre-trained language models. arXiv preprint arXiv:1908.10063 (2019)"},{"key":"3_CR9","unstructured":"Huang, K., Altosaar, J., Ranganath, R.: ClinicalBERT: modeling clinical notes and predicting hospital readmission. arXiv preprint arXiv:1904.05342 (2019)"},{"key":"3_CR10","unstructured":"Wu, S., et al.: BloombergGPT: a large language model for finance. arXiv preprint arXiv:2303.17564 (2023)"},{"key":"3_CR11","unstructured":"Driess, D., et al.: PaLM-E: an embodied multimodal language model. arXiv preprint arXiv:2303.03378 (2023)"},{"key":"3_CR12","unstructured":"Huang, S., et al.: Instruct2ACT: mapping multi-modality instructions to robotic actions with large language model. arXiv preprint arXiv:2305.11176 (2023)"},{"key":"3_CR13","first-page":"24824","volume":"35","author":"J Wei","year":"2022","unstructured":"Wei, J., et al.: Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural. Inf. Process. Syst. 35, 24824\u201324837 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3_CR14","unstructured":"Wang, X., et al.: Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171 (2022)"},{"key":"3_CR15","unstructured":"K\u0131c\u0131man, E., et al.: Causal reasoning and large language models: opening a new frontier for causality. arXiv preprint arXiv:2305.00050 (2023)"},{"key":"3_CR16","first-page":"22199","volume":"35","author":"T Kojima","year":"2022","unstructured":"Kojima, T., et al.: Large language models are zero-shot reasoners. Adv. Neural. Inf. Process. Syst. 35, 22199\u201322213 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3_CR17","doi-asserted-by":"crossref","unstructured":"Wan, X., et al.: Better zero-shot reasoning with self-adaptive prompting. arXiv preprint arXiv:2305.14106 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.216"},{"key":"3_CR18","unstructured":"Yao, S., et al.: ReAct: synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629 (2022)"},{"key":"3_CR19","unstructured":"Shinn, N., et al.: Reflexion: language agents with verbal reinforcement learning. Adv. Neural Inf. Process. Syst. 36 (2024)"},{"key":"3_CR20","doi-asserted-by":"crossref","unstructured":"Park, J.S., et al.: Generative agents: interactive simulacra of human behavior. In: Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology (2023)","DOI":"10.1145\/3586183.3606763"},{"key":"3_CR21","unstructured":"Wang, L., et al.: A survey on large language model based autonomous agents. arXiv preprint arXiv:2308.11432 (2023)"},{"key":"3_CR22","unstructured":"Xi, Z., et al.: The rise and potential of large language model based agents: a survey. arXiv preprint arXiv:2309.07864 (2023)"},{"key":"3_CR23","doi-asserted-by":"crossref","unstructured":"Wang, L., et al.: Plan-and-solve prompting: improving zero-shot chain-of-thought reasoning by large language models. arXiv preprint arXiv:2305.04091 (2023)","DOI":"10.18653\/v1\/2023.acl-long.147"},{"issue":"9","key":"3_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu, P., et al.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023)","journal-title":"ACM Comput. Surv."},{"key":"3_CR25","unstructured":"Sartor, G.: Legal reasoning. Treatise Legal Philos. Gen. Jurisprud. 5 (2005)"},{"issue":"4","key":"3_CR26","first-page":"100020","volume":"6","author":"BJ Jansen","year":"2022","unstructured":"Jansen, B.J., et al.: The illusion of data validity: why numbers about people are likely wrong. Data Inf. Manage. 6(4), 100020 (2022)","journal-title":"Data Inf. Manage."},{"key":"3_CR27","unstructured":"Chen, J., et al.: S-Agents: self-organizing agents in open-ended environment. arXiv preprint arXiv:2402.04578 (2024)"},{"key":"3_CR28","unstructured":"Zhuge, M., et al.: Mindstorms in natural language-based societies of mind. arXiv preprint arXiv:2305.17066 (2023)"},{"key":"3_CR29","unstructured":"Hao, R., et al.: ChatLLM network: More brains, more intelligence. arXiv preprint arXiv:2304.12998 (2023)"},{"key":"3_CR30","unstructured":"Liu, R., et al.: Training socially aligned language models in simulated human society. arXiv preprint arXiv:2305.16960 (2023)"},{"key":"3_CR31","unstructured":"Cai, T., et al.: Large language models as tool makers. arXiv preprint arXiv:2305.17126 (2023)"},{"key":"3_CR32","unstructured":"Yu, F., Quartey, L., Schilder, F.: Legal prompting: teaching a language model to think like a lawyer. arXiv preprint arXiv:2212.01326 (2022)"},{"key":"3_CR33","doi-asserted-by":"crossref","unstructured":"Jiang, C., Yang, X.: Legal syllogism prompting: teaching large language models for legal judgment prediction. In: Proceedings of the Nineteenth International Conference on Artificial Intelligence and Law (2023)","DOI":"10.1145\/3594536.3595170"},{"key":"3_CR34","doi-asserted-by":"crossref","unstructured":"Zhong, H., et al.: JEC-QA: a legal-domain question answering dataset. In: Proceedings of the AAAI Conference on Artificial Intelligence (2020)","DOI":"10.1609\/aaai.v34i05.6519"},{"key":"3_CR35","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"439","DOI":"10.1007\/978-3-030-32381-3_36","volume-title":"Chinese Computational Linguistics","author":"X Duan","year":"2019","unstructured":"Duan, X., et al.: CJRC: a reliable human-annotated benchmark dataset for chinese judicial reading comprehension. In: Sun, M., HUANG, X., Ji, H., Liu, Z., Liu, Y. (eds.) CCL 111112019. LNCS (LNAI), vol. 11856, pp. 439\u2013451. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32381-3_36"},{"key":"3_CR36","unstructured":"Huang, Q., et al.: Lawyer llama technical report. arXiv preprint arXiv:2305.15062 (2023)"},{"key":"3_CR37","unstructured":"Nguyen, H.-T.: A brief report on lawGPT 1.0: a virtual legal assistant based on GPT-3. arXiv preprint arXiv:2302.05729 (2023)"},{"key":"3_CR38","doi-asserted-by":"crossref","unstructured":"Hassanzadeh, T., Meybodi, M.R.: A new hybrid approach for data clustering using firefly algorithm and K-means. In: Proceedings of the 16th CSI International Symposium on Artificial Intelligence and Signal Processing (AISP 2012). IEEE (2012)","DOI":"10.1109\/AISP.2012.6313708"},{"key":"3_CR39","unstructured":"Yue, S., et al.: DISC-LawLLM: fine-tuning large language models for intelligent legal services. arXiv preprint arXiv:2309.11325 (2023)"},{"key":"3_CR40","unstructured":"Lu, J., et al.: Ziya-VL: bilingual large vision-language model via multi-task instruction tuning. arXiv preprint arXiv:2310.08166 (2023)"},{"key":"3_CR41","doi-asserted-by":"crossref","unstructured":"Du, Z., et al.: GLM: general language model pretraining with autoregressive blank infilling. arXiv preprint arXiv:2103.10360 (2021)","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"3_CR42","unstructured":"Yang, A., et al.: Baichuan 2: open large-scale language models. arXiv preprint arXiv:2309.10305 (2023)"},{"key":"3_CR43","unstructured":"Dai, Y., et al.: LAiW: a Chinese legal large language models benchmark (a technical report). arXiv preprint arXiv:2310.05620 (2023)"},{"key":"3_CR44","unstructured":"Cui, J., et al.: ChatLaw: open-source legal large language model with integrated external knowledge bases. arXiv preprint arXiv:2306.16092 (2023)"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-5678-0_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T16:26:03Z","timestamp":1722529563000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-5678-0_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819756773","9789819756780"],"references-count":44,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-5678-0_3","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 August 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tianjin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 August 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 August 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/2024\/index.htm","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}