{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,7]],"date-time":"2026-01-07T07:46:35Z","timestamp":1767771995518,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":42,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819794300"},{"type":"electronic","value":"9789819794317"}],"license":[{"start":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:00Z","timestamp":1730419200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:00Z","timestamp":1730419200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-9431-7_35","type":"book-chapter","created":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T14:03:04Z","timestamp":1730383384000},"page":"457-469","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Empowering LLMs for\u00a0Long-Text Information Extraction in\u00a0Chinese Legal Documents"],"prefix":"10.1007","author":[{"given":"Chenchen","family":"Shen","sequence":"first","affiliation":[]},{"given":"Chengwei","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Shengbin","family":"Yue","sequence":"additional","affiliation":[]},{"given":"Xiaoyu","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Yun","family":"Song","sequence":"additional","affiliation":[]},{"given":"Xuanjing","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Zhongyu","family":"Wei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,1]]},"reference":[{"key":"35_CR1","doi-asserted-by":"crossref","unstructured":"Akbik, A., Bergmann, T., Vollgraf, R.: Pooled contextualized embeddings for named entity recognition. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 724\u2013728 (2019)","DOI":"10.18653\/v1\/N19-1078"},{"key":"35_CR2","unstructured":"Bai, J., et al.: Qwen technical report. arXiv preprint arXiv:2309.16609 (2023)"},{"key":"35_CR3","unstructured":"Bai, Y., et al.: Longbench: a bilingual, multitask benchmark for long context understanding. arXiv preprint arXiv:2308.14508 (2023)"},{"key":"35_CR4","unstructured":"Beltagy, I., Peters, M.E., Cohan, A.: Longformer: the long-document transformer. arXiv preprint arXiv:2004.05150 (2020)"},{"key":"35_CR5","unstructured":"bloc97: Ntk-aware scaled rope allows llama models to have extended (8k+) context size without any fine-tuning and minimal perplexity degradation. https:\/\/www.reddit.com\/r\/LocalLLaMA\/comments\/14lz7j5\/ntkaware_scaled_rope_allows_llama_models_to_have\/"},{"key":"35_CR6","unstructured":"Bruckschen, M., et al.: Named entity recognition in the legal domain for ontology population. In: Workshop Programme, p.\u00a016 (2010)"},{"key":"35_CR7","unstructured":"Chen, S., Wong, S., Chen, L., Tian, Y.: Extending context window of large language models via positional interpolation. arXiv preprint arXiv:2306.15595 (2023)"},{"key":"35_CR8","doi-asserted-by":"crossref","unstructured":"Chen, Y., Xu, L., Liu, K., Zeng, D., Zhao, J.: Event extraction via dynamic multi-pooling convolutional neural networks. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 167\u2013176 (2015)","DOI":"10.3115\/v1\/P15-1017"},{"key":"35_CR9","unstructured":"Chen, Y., et al.: Longlora: efficient fine-tuning of long-context large language models. arXiv preprint arXiv:2309.12307 (2023)"},{"key":"35_CR10","unstructured":"Child, R., Gray, S., Radford, A., Sutskever, I.: Generating long sequences with sparse transformers. arXiv preprint arXiv:1904.10509 (2019)"},{"key":"35_CR11","doi-asserted-by":"crossref","unstructured":"Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.V., Salakhutdinov, R.: Transformer-xl: attentive language models beyond a fixed-length context. arXiv preprint arXiv:1901.02860 (2019)","DOI":"10.18653\/v1\/P19-1285"},{"key":"35_CR12","unstructured":"Dao, T.: Flashattention-2: faster attention with better parallelism and work partitioning. arXiv preprint arXiv:2307.08691 (2023)"},{"key":"35_CR13","doi-asserted-by":"crossref","unstructured":"Ding, S., et al.: Ernie-doc: a retrospective long-document modeling transformer. arXiv preprint arXiv:2012.15688 (2020)","DOI":"10.18653\/v1\/2021.acl-long.227"},{"key":"35_CR14","doi-asserted-by":"crossref","unstructured":"Du, Z., et al.: GLM: general language model pretraining with autoregressive blank infilling. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 320\u2013335 (2022)","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"35_CR15","doi-asserted-by":"crossref","unstructured":"Gui, H., Zhang, J., Ye, H., Zhang, N.: Instructie: a Chinese instruction-based information extraction dataset. arXiv preprint arXiv:2305.11527 (2023)","DOI":"10.1007\/978-3-031-77847-6_4"},{"key":"35_CR16","unstructured":"Hao, Y., Sun, Y., Dong, L., Han, Z., Gu, Y., Wei, F.: Structured prompting: scaling in-context learning to 1,000 examples. arXiv preprint arXiv:2212.06713 (2022)"},{"key":"35_CR17","unstructured":"Huang, Y., et al.: Advancing transformer architecture in long-context large language models: a comprehensive survey (2023)"},{"key":"35_CR18","doi-asserted-by":"crossref","unstructured":"Kwon, W., et al.: Efficient memory management for large language model serving with pagedattention. In: Proceedings of the 29th Symposium on Operating Systems Principles, pp. 611\u2013626 (2023)","DOI":"10.1145\/3600006.3613165"},{"key":"35_CR19","doi-asserted-by":"crossref","unstructured":"Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., Dyer, C.: Neural architectures for named entity recognition. arXiv preprint arXiv:1603.01360 (2016)","DOI":"10.18653\/v1\/N16-1030"},{"key":"35_CR20","doi-asserted-by":"publisher","first-page":"157","DOI":"10.1162\/tacl_a_00638","volume":"12","author":"NF Liu","year":"2024","unstructured":"Liu, N.F., et al.: Lost in the middle: how language models use long contexts. Trans. Assoc. Comput. Linguist. 12, 157\u2013173 (2024)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"35_CR21","unstructured":"Liu, X., Yan, H., Zhang, S., An, C., Qiu, X., Lin, D.: Scaling laws of rope-based extrapolation. arXiv preprint arXiv:2310.05209 (2023)"},{"key":"35_CR22","first-page":"2441","volume":"34","author":"X Ma","year":"2021","unstructured":"Ma, X., et al.: Luna: linear unified nested attention. Adv. Neural. Inf. Process. Syst. 34, 2441\u20132453 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"35_CR23","unstructured":"Masala, M., Rebedea, T., Velicu, H.: Improving legal judgement prediction in Romanian with long text encoders. In: Melero, M., Sakti, S., Soria, C. (eds.) Proceedings of the 3rd Annual Meeting of the Special Interest Group on Under-resourced Languages @ LREC-COLING 2024, pp. 126\u2013132. ELRA and ICCL, Torino, Italia (2024). https:\/\/aclanthology.org\/2024.sigul-1.16"},{"key":"35_CR24","doi-asserted-by":"crossref","unstructured":"Miwa, M., Bansal, M.: End-to-end relation extraction using LSTMs on sequences and tree structures. arXiv preprint arXiv:1601.00770 (2016)","DOI":"10.18653\/v1\/P16-1105"},{"key":"35_CR25","doi-asserted-by":"crossref","unstructured":"Nguyen, T.H., Cho, K., Grishman, R.: Joint event extraction via recurrent neural networks. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 300\u2013309 (2016)","DOI":"10.18653\/v1\/N16-1034"},{"key":"35_CR26","unstructured":"Ratner, N., et al.: Parallel context windows for large language models. arXiv preprint arXiv:2212.10947 (2022)"},{"key":"35_CR27","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1162\/tacl_a_00353","volume":"9","author":"A Roy","year":"2021","unstructured":"Roy, A., Saffar, M., Vaswani, A., Grangier, D.: Efficient content-based sparse attention with routing transformers. Trans. Assoc. Comput. Linguist. 9, 53\u201368 (2021)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"35_CR28","unstructured":"Su, J.: Upgrade road of transformer:10.rope is a $$\\beta $$-base encoding. https:\/\/spaces.ac.cn\/archives\/9675"},{"key":"35_CR29","doi-asserted-by":"crossref","unstructured":"Su, J., Lu, Y., Pan, S., Murtadha, A., Wen, B., Liu, Y.: Roformer: enhanced transformer with rotary position embedding (2023)","DOI":"10.1016\/j.neucom.2023.127063"},{"key":"35_CR30","doi-asserted-by":"crossref","unstructured":"Sukhbaatar, S., Grave, E., Bojanowski, P., Joulin, A.: Adaptive attention span in transformers. arXiv preprint arXiv:1905.07799 (2019)","DOI":"10.18653\/v1\/P19-1032"},{"issue":"2","key":"35_CR31","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.clsr.2014.01.009","volume":"30","author":"M Truyens","year":"2014","unstructured":"Truyens, M., Van Eecke, P.: Legal aspects of text mining. Comput. Law Secur. Rev. 30(2), 153\u2013170 (2014)","journal-title":"Comput. Law Secur. Rev."},{"key":"35_CR32","doi-asserted-by":"publisher","unstructured":"Tuteja, M., Gonz\u00e1lez\u00a0Jucl\u00e0, D.: Long text classification using transformers with paragraph selection strategies. In: Preotiuc-Pietro, D., Goanta, C., Chalkidis, I., Barrett, L., Spanakis, G., Aletras, N. (eds.) Proceedings of the Natural Legal Language Processing Workshop 2023, pp. 17\u201324. Association for Computational Linguistics, Singapore (2023). https:\/\/doi.org\/10.18653\/v1\/2023.nllp-1.3, https:\/\/aclanthology.org\/2023.nllp-1.3","DOI":"10.18653\/v1\/2023.nllp-1.3"},{"key":"35_CR33","doi-asserted-by":"crossref","unstructured":"Vacek, T., Teo, R., Song, D., Nugent, T., Cowling, C., Schilder, F.: Litigation analytics: Case outcomes extracted from us federal court dockets. In: Proceedings of the Natural Legal Language Processing Workshop 2019, pp. 45\u201354 (2019)","DOI":"10.18653\/v1\/W19-2206"},{"key":"35_CR34","unstructured":"Vaswani, A., et al.: Attention is all you need. CoRR abs\/1706.03762 (2017). http:\/\/arxiv.org\/abs\/1706.03762"},{"key":"35_CR35","unstructured":"Wang, S., Li, B.Z., Khabsa, M., Fang, H., Ma, H.: Linformer: self-attention with linear complexity. arXiv preprint arXiv:2006.04768 (2020)"},{"key":"35_CR36","unstructured":"Wang, X., et\u00a0al.: Instructuie: multi-task instruction tuning for unified information extraction. arXiv preprint arXiv:2304.08085 (2023)"},{"key":"35_CR37","doi-asserted-by":"publisher","unstructured":"Xie, S., Li, L., Yuan, J., Xie, Q., Tao, X.: Long legal article question answering via cascaded key segment learning (student abstract). In: Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence and Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence and Thirteenth Symposium on Educational Advances in Artificial Intelligence. AAAI\u201923\/IAAI\u201923\/EAAI\u201923, AAAI Press (2023). https:\/\/doi.org\/10.1609\/aaai.v37i13.27042, https:\/\/doi.org\/10.1609\/aaai.v37i13.27042","DOI":"10.1609\/aaai.v37i13.27042"},{"key":"35_CR38","doi-asserted-by":"crossref","unstructured":"Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., Hovy, E.: Hierarchical attention networks for document classification. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1480\u20131489 (2016)","DOI":"10.18653\/v1\/N16-1174"},{"key":"35_CR39","unstructured":"Yin, X., Zheng, D., Lu, Z., Liu, R.: Neural entity reasoner for global consistency in NER. arXiv preprint arXiv:1810.00347 (2018)"},{"key":"35_CR40","unstructured":"Yue, S., et al.: Disc-lawllm: fine-tuning large language models for intelligent legal services (2023)"},{"key":"35_CR41","doi-asserted-by":"crossref","unstructured":"Zeng, D., Liu, K., Chen, Y., Zhao, J.: Distant supervision for relation extraction via piecewise convolutional neural networks. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1753\u20131762 (2015)","DOI":"10.18653\/v1\/D15-1203"},{"key":"35_CR42","doi-asserted-by":"publisher","first-page":"18250","DOI":"10.1109\/ACCESS.2017.2745208","volume":"5","author":"N Zhang","year":"2017","unstructured":"Zhang, N., Pu, Y.F., Yang, S.Q., Zhou, J.L., Gao, J.K.: An ontological Chinese legal consultation system. IEEE Access 5, 18250\u201318261 (2017)","journal-title":"IEEE Access"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Chinese Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-9431-7_35","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T15:54:51Z","timestamp":1732982091000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-9431-7_35"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,1]]},"ISBN":["9789819794300","9789819794317"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-9431-7_35","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,1]]},"assertion":[{"value":"1 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLPCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF International Conference on Natural Language Processing and Chinese Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nlpcc2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/tcci.ccf.org.cn\/conference\/2024\/index.php","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}