{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T17:19:36Z","timestamp":1776100776263,"version":"3.50.1"},"reference-count":66,"publisher":"Association for Computing Machinery (ACM)","issue":"2","license":[{"start":{"date-parts":[[2023,11,8]],"date-time":"2023-11-08T00:00:00Z","timestamp":1699401600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62002194"],"award-info":[{"award-number":["62002194"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Inf. Syst."],"published-print":{"date-parts":[[2024,3,31]]},"abstract":"<jats:p>Legal case retrieval has received increasing attention in recent years. However, compared to ad hoc retrieval tasks, legal case retrieval has its unique challenges. First, case documents are rather lengthy and contain complex legal structures. Therefore, it is difficult for most existing dense retrieval models to encode an entire document and capture its inherent complex structure information. Most existing methods simply truncate part of the document content to meet the input length limit of PLMs, which will lead to information loss. Additionally, the definition of relevance in the legal domain differs from that in the general domain. Previous semantic-based or lexical-based methods fail to provide a comprehensive understanding of the relevance of legal cases. In this article, we propose a Structured Legal case Retrieval (SLR) framework, which incorporates internal and external structural information to address the above two challenges. Specifically, to avoid the truncation of long legal documents, the internal structural information, which is the organization pattern of legal documents, can be utilized to split a case document into segments. By dividing the document-level semantic matching task into segment-level subtasks, SLR can separately process segments using different methods based on the characteristic of each segment. In this way, the key elements of a case document can be highlighted without losing other content information. Second, toward a better understanding of relevance in the legal domain, we investigate the connections between criminal charges appearing in large-scale case corpus to generate a chargewise relation graph. Then, the similarity between criminal charges can be pre-computed as the external structural information to enhance the recognition of relevant cases. Finally, a learning-to-rank algorithm integrates the features collected from internal and external structures to output the final retrieval results. Experimental results on public legal case retrieval benchmarks demonstrate the superior effectiveness of SLR over existing state-of-the-art baselines, including traditional bag-of-words and neural-based methods. Furthermore, we conduct a case study to visualize how the proposed model focuses on key elements and improves retrieval performance.<\/jats:p>","DOI":"10.1145\/3609796","type":"journal-article","created":{"date-parts":[[2023,7,19]],"date-time":"2023-07-19T12:10:06Z","timestamp":1689768606000},"page":"1-28","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":11,"title":["Incorporating Structural Information into Legal Case Retrieval"],"prefix":"10.1145","volume":"42","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2107-7408","authenticated-orcid":false,"given":"Yixiao","family":"Ma","sequence":"first","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2539-8954","authenticated-orcid":false,"given":"Yueyue","family":"Wu","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5030-709X","authenticated-orcid":false,"given":"Qingyao","family":"Ai","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0140-4512","authenticated-orcid":false,"given":"Yiqun","family":"Liu","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1727-8311","authenticated-orcid":false,"given":"Yunqiu","family":"Shao","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3158-1920","authenticated-orcid":false,"given":"Min","family":"Zhang","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8762-8268","authenticated-orcid":false,"given":"Shaoping","family":"Ma","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Institute for Internet Judiciary, Tsinghua University. Quan Cheng Laboratory, China"}]}],"member":"320","published-online":{"date-parts":[[2023,11,8]]},"reference":[{"key":"e_1_3_2_2_2","first-page":"162","volume-title":"Proceedings of the 2nd International Conference on Design of Experimental Search & Information REtrieval Systems","author":"Askari A. A.","year":"2021","unstructured":"A. A. Askari, S. V. Verberne, O. Alonso, S. Marchesin, M. Najork, and G. Silvello. 2021. Combining lexical and neural retrieval with longformer-based summarization for effective case law retrieva. In Proceedings of the 2nd International Conference on Design of Experimental Search & Information REtrieval Systems. CEUR, 162\u2013170."},{"key":"e_1_3_2_3_2","article-title":"Longformer: The long-document transformer","author":"Beltagy Iz","year":"2020","unstructured":"Iz Beltagy, Matthew E. Peters, and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv:2004.05150. Retrieved from https:\/\/arxiv.org\/abs\/2004.05150","journal-title":"arXiv:2004.05150"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10506-012-9131-x"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.103069"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1023\/A:1010933404324"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-30490-4_20"},{"key":"e_1_3_2_8_2","article-title":"Generating long sequences with sparse transformers","author":"Child Rewon","year":"2019","unstructured":"Rewon Child, Scott Gray, Alec Radford, and Ilya Sutskever. 2019. Generating long sequences with sparse transformers. arXiv:1904.10509. Retrieved from https:\/\/arxiv.org\/abs\/1904.10509","journal-title":"arXiv:1904.10509"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1080\/01621459.1998.10473750"},{"key":"e_1_3_2_10_2","article-title":"Pre-training with whole word masking for chinese bert","author":"Cui Yiming","year":"2019","unstructured":"Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, and Guoping Hu. 2019. Pre-training with whole word masking for chinese bert. arXiv:1906.08101. Retrieved from https:\/\/arxiv.org\/abs\/1906.08101","journal-title":"arXiv:1906.08101"},{"key":"e_1_3_2_11_2","article-title":"Transformer-xl: Attentive language models beyond a fixed-length context","author":"Dai Zihang","year":"2019","unstructured":"Zihang Dai, Zhilin Yang, Yiming Yang, Jaime Carbonell, Quoc V Le, and Ruslan Salakhutdinov. 2019. Transformer-xl: Attentive language models beyond a fixed-length context. arXiv:1901.02860. Retrieved from https:\/\/arxiv.org\/abs\/1901.02860","journal-title":"arXiv:1901.02860"},{"key":"e_1_3_2_12_2","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805. Retrieved from https:\/\/arxiv.org\/abs\/1810.04805","journal-title":"arXiv:1810.04805"},{"key":"e_1_3_2_13_2","article-title":"Inheritance-guided hierarchical assignment for clinical automatic diagnosis","author":"Du Yichao","year":"2021","unstructured":"Yichao Du, Pengfei Luo, Xudong Hong, Tong Xu, Zhe Zhang, Chao Ren, Yi Zheng, and Enhong Chen. 2021. Inheritance-guided hierarchical assignment for clinical automatic diagnosis. arXiv:2101.11374. Retrieved from https:\/\/arxiv.org\/abs\/2101.11374","journal-title":"arXiv:2101.11374"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-32695-0_6"},{"key":"e_1_3_2_15_2","article-title":"Learning fine-grained fact-article correspondence in legal cases","author":"Ge Jidong","year":"2021","unstructured":"Jidong Ge, Xiaoyu Shen, Chuanyi Li, Wei Hu, Bin Luo, et\u00a0al. 2021. Learning fine-grained fact-article correspondence in legal cases. arXiv:2104.10726. Retrieved from https:\/\/arxiv.org\/abs\/2104.10726","journal-title":"arXiv:2104.10726"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939754"},{"key":"e_1_3_2_17_2","article-title":"Don\u2019t stop pretraining: Adapt language models to domains and tasks","author":"Gururangan Suchin","year":"2020","unstructured":"Suchin Gururangan, Ana Marasovi\u0107, Swabha Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, and Noah A Smith. 2020. Don\u2019t stop pretraining: Adapt language models to domains and tasks. arXiv:2004.10964. Retrieved from https:\/\/arxiv.org\/abs\/2004.10964","journal-title":"arXiv:2004.10964"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1111\/jels.12230"},{"key":"e_1_3_2_19_2","first-page":"1","article-title":"Investigating sentence severity with judicial open data: A case study on sentencing high-tech crime in the Dutch criminal justice system","author":"Hartel Pieter","year":"2022","unstructured":"Pieter Hartel, Rolf + Wegberg, and Mark van Staalduinen. 2022. Investigating sentence severity with judicial open data: A case study on sentencing high-tech crime in the Dutch criminal justice system. Eur. J. Crim. Policy Res. (2022), 1\u201321.","journal-title":"Eur. J. Crim. Policy Res."},{"key":"e_1_3_2_20_2","volume-title":"Proceedings of the COLIEE Workshop in ICAIL","author":"Juliano Rabelo","year":"2021","unstructured":"Rabelo Juliano, Goebel Randy, Kano Yoshinobu, Kim Mi-Young, Yoshioka Masaharu, and Satoh Ken. 2021. Summary of the competition on legal information extraction\/entailment (coliee) 2021. In Proceedings of the COLIEE Workshop in ICAIL."},{"key":"e_1_3_2_21_2","first-page":"3146","article-title":"Lightgbm: A highly efficient gradient boosting decision tree","volume":"30","author":"Ke Guolin","year":"2017","unstructured":"Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 30 (2017), 3146\u20133154.","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"e_1_3_2_22_2","article-title":"Adam: A method for stochastic optimization","author":"Kingma Diederik P.","year":"2014","unstructured":"Diederik P. Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv:1412.6980. Retrieved from https:\/\/arxiv.org\/abs\/1412.6980","journal-title":"arXiv:1412.6980"},{"key":"e_1_3_2_23_2","article-title":"Semi-supervised classification with graph convolutional networks","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv:1609.02907. Retrieved from https:\/\/arxiv.org\/abs\/1609.02907","journal-title":"arXiv:1609.02907"},{"key":"e_1_3_2_24_2","article-title":"SAILER: Structure-aware pre-trained language model for legal case retrieval","author":"Li Haitao","year":"2023","unstructured":"Haitao Li, Qingyao Ai, Jia Chen, Qian Dong, Yueyue Wu, Yiqun Liu, Chong Chen, and Qi Tian. 2023. SAILER: Structure-aware pre-trained language model for legal case retrieval. arXiv:2304.11370. Retrieved from https:\/\/arxiv.org\/abs\/2304.11370","journal-title":"arXiv:2304.11370"},{"key":"e_1_3_2_25_2","unstructured":"Haitao Li Weihang Su Changyue Wang Yueyue Wu Qingyao Ai and Yiqun Liu. 2023. THUIR@COLIEE 2023: Incorporating structural knowledge into pre-trained language models for legal case retrieval. arxiv:2305.06812 [cs.IR]. Retrieved from https:\/\/arxiv.org\/abs\/2305.06812"},{"issue":"3","key":"e_1_3_2_26_2","first-page":"1","article-title":"The power of selecting key blocks with local pre-ranking for long document information retrieval","volume":"41","author":"Li Minghan","year":"2023","unstructured":"Minghan Li, Diana Nicoleta Popa, Johan Chagnon, Yagmur Gizem Cinar, and Eric Gaussier. 2023. The power of selecting key blocks with local pre-ranking for long document information retrieval. ACM Trans. Inf. Syst. 41, 3 (2023), 1\u201335.","journal-title":"ACM Trans. Inf. Syst."},{"key":"e_1_3_2_27_2","doi-asserted-by":"crossref","unstructured":"Tie-Yan Liu and others. 2009. Learning to rank for information retrieval. Foundations and Trends\u00ae in Information Retrieval 3 3 (2009) 225\u2013331.","DOI":"10.1561\/1500000016"},{"key":"e_1_3_2_28_2","article-title":"Roberta: A robustly optimized bert pretraining approach","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692. Retrieved from https:\/\/arxiv.org\/abs\/1907.11692","journal-title":"arXiv:1907.11692"},{"issue":"3","key":"e_1_3_2_29_2","first-page":"62","article-title":"MLPV: Text representation of scientific papers based on structural information and Doc2vec","volume":"3","author":"Lu Yonghe","year":"2019","unstructured":"Yonghe Lu, Yuanyuan Zhai, Jiayi Luo, and Yongshan Chen. 2019. MLPV: Text representation of scientific papers based on structural information and Doc2vec. Am. J. Inf. Sci. Technol. 3, 3 (2019), 62\u201371.","journal-title":"Am. J. Inf. Sci. Technol."},{"key":"e_1_3_2_30_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2021.102780"},{"key":"e_1_3_2_31_2","unstructured":"Yixiao Ma Yunqiu Shao Bulou Liu Yiqun Liu Min Zhang and Shaoping Ma. 2021. Retrieving legal cases from a large-scale candidate corpus. Proceedings of the Eighth International Competition on Legal Information Extraction\/Entailment (COLIEE\u201921) (2021)."},{"key":"e_1_3_2_32_2","first-page":"22","article-title":"LeCaRD: A legal case retrieval dataset for chinese law system","volume":"2","author":"Ma Yixiao","year":"2021","unstructured":"Yixiao Ma, Yunqiu Shao, Yueyue Wu, Yiqun Liu, Ruizhe Zhang, Min Zhang, and Shaoping Ma. 2021. LeCaRD: A legal case retrieval dataset for chinese law system. Inf. Retriev. 2 (2021), 22.","journal-title":"Inf. Retriev."},{"key":"e_1_3_2_33_2","article-title":"Efficient estimation of word representations in vector space","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. arXiv:1301.3781. Retrieved from https:\/\/arxiv.org\/abs\/1301.3781","journal-title":"arXiv:1301.3781"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1145\/2740908.2744717"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1038\/nbt1206-1565"},{"key":"e_1_3_2_36_2","first-page":"8026","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume":"32","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et\u00a0al. 2019. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 32 (2019), 8026\u20138037.","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1145\/290941.291008"},{"key":"e_1_3_2_38_2","first-page":"34","volume-title":"JSAI International Symposium on Artificial Intelligence","author":"Rabelo Juliano","year":"2019","unstructured":"Juliano Rabelo, Mi-Young Kim, Randy Goebel, Masaharu Yoshioka, Yoshinobu Kano, and Ken Satoh. 2019. A summary of the COLIEE 2019 competition. In JSAI International Symposium on Artificial Intelligence. Springer, 34\u201349."},{"key":"e_1_3_2_39_2","first-page":"196","volume-title":"JSAI International Symposium on Artificial Intelligence","author":"Rabelo Juliano","year":"2020","unstructured":"Juliano Rabelo, Mi-Young Kim, Randy Goebel, Masaharu Yoshioka, Yoshinobu Kano, and Ken Satoh. 2020. COLIEE 2020: Methods for legal document retrieval and entailment. In JSAI International Symposium on Artificial Intelligence. Springer, 196\u2013210."},{"key":"e_1_3_2_40_2","unstructured":"Radim \u0158eh\u016f\u0159ek Petr Sojka and others. 2011. Gensim\u2014statistical semantics in python. Retrieved from genism.org (2011)."},{"key":"e_1_3_2_41_2","first-page":"109","article-title":"Okapi at TREC-3","volume":"109","author":"Robertson Stephen E.","year":"1995","unstructured":"Stephen E. Robertson, Steve Walker, Susan Jones, Micheline M. Hancock-Beaulieu, Mike Gatford, et\u00a0al. 1995. Okapi at TREC-3. Nist Special Publication Sp 109 (1995), 109.","journal-title":"Nist Special Publication Sp"},{"key":"e_1_3_2_42_2","doi-asserted-by":"publisher","DOI":"10.1016\/0306-4573(88)90021-0"},{"key":"e_1_3_2_43_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10506-009-9075-y"},{"key":"e_1_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.1145\/3462757.3466149"},{"key":"e_1_3_2_45_2","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/484"},{"key":"e_1_3_2_46_2","article-title":"Understanding relevance judgments in legal case retrieval","author":"Shao Yunqiu","year":"2022","unstructured":"Yunqiu Shao, Yueyue Wu, Yiqun Liu, Jiaxin Mao, and Shaoping Ma. 2022. Understanding relevance judgments in legal case retrieval. ACM Trans. Inf. Syst. (2022).","journal-title":"ACM Trans. Inf. Syst."},{"key":"e_1_3_2_47_2","article-title":"Adaptive attention span in transformers","author":"Sukhbaatar Sainbayar","year":"2019","unstructured":"Sainbayar Sukhbaatar, Edouard Grave, Piotr Bojanowski, and Armand Joulin. 2019. Adaptive attention span in transformers. arXiv:1905.07799. Retrieved from https:\/\/arxiv.org\/abs\/1905.07799","journal-title":"arXiv:1905.07799"},{"key":"e_1_3_2_48_2","unstructured":"J. Sun. 2012. Jieba chinese word segmentation tool. https:\/\/github.com\/fxsjy\/jieba"},{"key":"e_1_3_2_49_2","article-title":"Law article-enhanced legal case matching: A model-agnostic causal learning approach","author":"Sun Zhongxiang","year":"2022","unstructured":"Zhongxiang Sun, Jun Xu, Xiao Zhang, Zhenhua Dong, and Ji-Rong Wen. 2022. Law article-enhanced legal case matching: A model-agnostic causal learning approach. arXiv:2210.11012. Retrieved from https:\/\/arxiv.org\/abs\/2210.11012","journal-title":"arXiv:2210.11012"},{"key":"e_1_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10506-020-09262-4"},{"key":"e_1_3_2_51_2","doi-asserted-by":"publisher","DOI":"10.1145\/3322640.3326740"},{"key":"e_1_3_2_52_2","first-page":"167","volume-title":"Legal Knowledge and Information Systems","author":"Trompper Maarten","year":"2016","unstructured":"Maarten Trompper and Radboud Winkels. 2016. Automatic assignment of section structure to texts of Dutch court judgments. In Legal Knowledge and Information Systems. IOS Press, 167\u2013172."},{"key":"e_1_3_2_53_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10506-017-9195-8"},{"key":"e_1_3_2_54_2","first-page":"5998","volume-title":"Advances in Neural Information Processing Systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems. 5998\u20136008."},{"key":"e_1_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1145\/3331184.3331223"},{"key":"e_1_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462873"},{"key":"e_1_3_2_57_2","first-page":"269","volume-title":"JSAI International Symposium on Artificial Intelligence","author":"Westermann Hannes","year":"2020","unstructured":"Hannes Westermann, Jaromir Savelka, and Karim Benyekhlef. 2020. Paragraph similarity scoring and fine-tuned BERT for legal information retrieval and entailment. In JSAI International Symposium on Artificial Intelligence. Springer, 269\u2013285."},{"key":"e_1_3_2_58_2","article-title":"Huggingface\u2019s transformers: State-of-the-art natural language processing","author":"Wolf Thomas","year":"2019","unstructured":"Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R\u00e9mi Louf, Morgan Funtowicz, et\u00a0al. 2019. Huggingface\u2019s transformers: State-of-the-art natural language processing. arXiv:1910.03771. Retrieved from https:\/\/arxiv.org\/abs\/1910.03771","journal-title":"arXiv:1910.03771"},{"key":"e_1_3_2_59_2","doi-asserted-by":"crossref","unstructured":"Chaojun Xiao Xueyu Hu Zhiyuan Liu Cunchao Tu and Maosong Sun. 2021. Lawformer: A pre-trained language model for chinese legal long documents. AI Open 2 (2021) 79\u201384.","DOI":"10.1016\/j.aiopen.2021.06.003"},{"key":"e_1_3_2_60_2","article-title":"Cail2018: A large-scale legal dataset for judgment prediction","author":"Xiao Chaojun","year":"2018","unstructured":"Chaojun Xiao, Haoxi Zhong, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Yansong Feng, Xianpei Han, Zhen Hu, Heng Wang, et\u00a0al. 2018. Cail2018: A large-scale legal dataset for judgment prediction. arXiv:1807.02478. Retrieved from https:\/\/arxiv.org\/abs\/1807.02478","journal-title":"arXiv:1807.02478"},{"key":"e_1_3_2_61_2","doi-asserted-by":"publisher","DOI":"10.1145\/1645953.1646295"},{"key":"e_1_3_2_62_2","doi-asserted-by":"publisher","DOI":"10.1145\/3450626.3459873"},{"key":"e_1_3_2_63_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462826"},{"key":"e_1_3_2_64_2","doi-asserted-by":"publisher","DOI":"10.1007\/11552413_49"},{"key":"e_1_3_2_65_2","doi-asserted-by":"publisher","DOI":"10.1186\/s40649-019-0069-y"},{"key":"e_1_3_2_66_2","volume-title":"Open Chinese Language Pre-trained Model Zoo","author":"Zhong Haoxi","year":"2019","unstructured":"Haoxi Zhong, Zhengyan Zhang, Zhiyuan Liu, and Maosong Sun. 2019. Open Chinese Language Pre-trained Model Zoo. Technical Report."},{"key":"e_1_3_2_67_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2021.01.001"}],"container-title":["ACM Transactions on Information Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3609796","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3609796","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:38:01Z","timestamp":1750178281000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3609796"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,8]]},"references-count":66,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,3,31]]}},"alternative-id":["10.1145\/3609796"],"URL":"https:\/\/doi.org\/10.1145\/3609796","relation":{},"ISSN":["1046-8188","1558-2868"],"issn-type":[{"value":"1046-8188","type":"print"},{"value":"1558-2868","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,11,8]]},"assertion":[{"value":"2023-01-04","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-06-12","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-11-08","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}