{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T14:18:42Z","timestamp":1774448322003,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":51,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,6,21]],"date-time":"2021-06-21T00:00:00Z","timestamp":1624233600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,6,21]]},"DOI":"10.1145\/3462757.3466103","type":"proceedings-article","created":{"date-parts":[[2021,7,28]],"date-time":"2021-07-28T06:50:46Z","timestamp":1627455046000},"page":"295-300","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["To tune or not to tune?"],"prefix":"10.1145","author":[{"given":"Guilherme Moraes","family":"Rosa","sequence":"first","affiliation":[{"name":"NeuralMind, Brazil and University of Campinas (Unicamp), Brazil"}]},{"given":"Ruan Chaves","family":"Rodrigues","sequence":"additional","affiliation":[{"name":"NeuralMind, Brazil and Federal University of Goi\u00e1s (UFG), Brazil"}]},{"given":"Roberto","family":"de Alencar Lotufo","sequence":"additional","affiliation":[{"name":"NeuralMind, Brazil and University of Campinas (Unicamp), Brazil"}]},{"given":"Rodrigo","family":"Nogueira","sequence":"additional","affiliation":[{"name":"NeuralMind, Brazil and University of Campinas (Unicamp), Brazil and University of Waterloo, Canada"}]}],"member":"320","published-online":{"date-parts":[[2021,7,27]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. arXiv:1611.09268v3","author":"Bajaj Payal","year":"2018","unstructured":"Payal Bajaj , Daniel Campos , Nick Craswell , Li Deng , Jianfeng Gao , Xiaodong Liu , Rangan Majumder , Andrew McNamara , Bhaskar Mitra , Tri Nguyen , Mir Rosenberg , Xia Song , Alina Stoica , Saurabh Tiwary , and Tong Wang . 2018 . MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. arXiv:1611.09268v3 (2018). Payal Bajaj, Daniel Campos, Nick Craswell, Li Deng, Jianfeng Gao, Xiaodong Liu, Rangan Majumder, Andrew McNamara, Bhaskar Mitra, Tri Nguyen, Mir Rosenberg, Xia Song, Alina Stoica, Saurabh Tiwary, and Tong Wang. 2018. MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. arXiv:1611.09268v3 (2018)."},{"key":"e_1_3_2_1_2_1","volume-title":"LegalDB: Long DistilBERT for Legal Document Classification. In 2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT). IEEE, 1--4.","author":"Bambroo Purbid","year":"2021","unstructured":"Purbid Bambroo and Aditi Awasthi . 2021 . LegalDB: Long DistilBERT for Legal Document Classification. In 2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT). IEEE, 1--4. Purbid Bambroo and Aditi Awasthi. 2021. LegalDB: Long DistilBERT for Legal Document Classification. In 2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT). IEEE, 1--4."},{"key":"e_1_3_2_1_3_1","unstructured":"Hangbo Bao Li Dong Furu Wei Wenhui Wang Nan Yang Xiaodong Liu Yu Wang Songhao Piao Jianfeng Gao Ming Zhou and Hsiao-Wuen Hon. 2020. UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training. ArXiv. https:\/\/www.microsoft.com\/en-us\/research\/publication\/unilmv2-pseudo-masked-language-models-for-unified-language-model-pre-training\/  Hangbo Bao Li Dong Furu Wei Wenhui Wang Nan Yang Xiaodong Liu Yu Wang Songhao Piao Jianfeng Gao Ming Zhou and Hsiao-Wuen Hon. 2020. UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training. ArXiv. https:\/\/www.microsoft.com\/en-us\/research\/publication\/unilmv2-pseudo-masked-language-models-for-unified-language-model-pre-training\/"},{"key":"e_1_3_2_1_4_1","volume-title":"Lin (Eds.)","volume":"33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown , Benjamin Mann , Nick Ryder , Melanie Subbiah , Jared D Kaplan , Prafulla Dhariwal , Arvind Neelakantan , Pranav Shyam , Girish Sastry , Amanda Askell , Sandhini Agarwal , Ariel Herbert-Voss , Gretchen Krueger , Tom Henighan , Rewon Child , Aditya Ramesh , Daniel Ziegler , Jeffrey Wu , Clemens Winter , Chris Hesse , Mark Chen , Eric Sigler , Mateusz Litwin , Scott Gray , Benjamin Chess , Jack Clark , Christopher Berner , Sam McCandlish , Alec Radford , Ilya Sutskever , and Dario Amodei . 2020 . Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H . Lin (Eds.) , Vol. 33 . Curran Associates, Inc. , 1877--1901. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 1877--1901. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf"},{"key":"e_1_3_2_1_5_1","first-page":"171","volume-title":"Deep learning in law: early adaptation and legal word embeddings trained on large corpora. Artificial Intelligence and Law","author":"Chalkidis Ilias","year":"2019","unstructured":"Ilias Chalkidis and Dimitrios Kampas . 2019. Deep learning in law: early adaptation and legal word embeddings trained on large corpora. Artificial Intelligence and Law volume 27 , pages 171 -- 198 ( 2019 ) (2019). Ilias Chalkidis and Dimitrios Kampas. 2019. Deep learning in law: early adaptation and legal word embeddings trained on large corpora. Artificial Intelligence and Law volume 27, pages 171--198(2019) (2019)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Sanyuan Chen Yutai Hou Yiming Cui Wanxiang Che Ting Liu and Xiangzhan Yu. 2020. Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting. arXiv:2004.12651 [cs.CL]  Sanyuan Chen Yutai Hou Yiming Cui Wanxiang Che Ting Liu and Xiangzhan Yu. 2020. Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting. arXiv:2004.12651 [cs.CL]","DOI":"10.18653\/v1\/2020.emnlp-main.634"},{"key":"e_1_3_2_1_7_1","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","volume":"1","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding . In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies , Volume 1 (Long and Short Papers). 4171--4186. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 4171--4186."},{"key":"e_1_3_2_1_8_1","first-page":"9","volume-title":"Summarization and Multi-Label Classification. AICCC '18: Proceedings of the 2018 Artificial Intelligence and Cloud Computing Conference","author":"Elnaggar Ahmed","year":"2018","unstructured":"Ahmed Elnaggar , Christoph Gebendorfer , Ingo Glaser , and Florian Matthes . 2018 . Multi-Task Deep Learning for Legal Document Translation , Summarization and Multi-Label Classification. AICCC '18: Proceedings of the 2018 Artificial Intelligence and Cloud Computing Conference December 2018 Pages 9 -- 15 (2018). Ahmed Elnaggar, Christoph Gebendorfer, Ingo Glaser, and Florian Matthes. 2018. Multi-Task Deep Learning for Legal Document Translation, Summarization and Multi-Label Classification. AICCC '18: Proceedings of the 2018 Artificial Intelligence and Cloud Computing Conference December 2018 Pages 9--15 (2018)."},{"key":"e_1_3_2_1_9_1","first-page":"41","volume-title":"Stop Illegal Comments: A Multi-Task Deep Learning Approach. AICCC '18: Proceedings of the 2018 Artificial Intelligence and Cloud Computing Conference","author":"Elnaggar Ahmed","year":"2018","unstructured":"Ahmed Elnaggar , Bernhard Waltl , Ingo Glaser , J\u00f6rg Landthaler , Elena Scepankova , and Florian Matthes . 2018 . Stop Illegal Comments: A Multi-Task Deep Learning Approach. AICCC '18: Proceedings of the 2018 Artificial Intelligence and Cloud Computing Conference December 2018 Pages 41 -- 47 (2018). Ahmed Elnaggar, Bernhard Waltl, Ingo Glaser, J\u00f6rg Landthaler, Elena Scepankova, and Florian Matthes. 2018. Stop Illegal Comments: A Multi-Task Deep Learning Approach. AICCC '18: Proceedings of the 2018 Artificial Intelligence and Cloud Computing Conference December 2018 Pages 41--47 (2018)."},{"key":"e_1_3_2_1_10_1","volume-title":"Workshop on Document Intelligence at NeurIPS","author":"Elwany Emad","year":"2019","unstructured":"Emad Elwany , Dave Moore , and Gaurav Oberoi . 2019 . BERT Goes to Law School: Quantifying the Competitive Advantage of Access to Large Legal Corpora in Contract Understanding . In Workshop on Document Intelligence at NeurIPS 2019. Emad Elwany, Dave Moore, and Gaurav Oberoi. 2019. BERT Goes to Law School: Quantifying the Competitive Advantage of Access to Large Legal Corpora in Contract Understanding. In Workshop on Document Intelligence at NeurIPS 2019."},{"key":"e_1_3_2_1_11_1","volume-title":"Rethink Training of BERT Rerankers in Multi-Stage Retrieval Pipeline. arXiv preprint arXiv:2101.08751","author":"Gao Luyu","year":"2021","unstructured":"Luyu Gao , Zhuyun Dai , and Jamie Callan . 2021. Rethink Training of BERT Rerankers in Multi-Stage Retrieval Pipeline. arXiv preprint arXiv:2101.08751 ( 2021 ). Luyu Gao, Zhuyun Dai, and Jamie Callan. 2021. Rethink Training of BERT Rerankers in Multi-Stage Retrieval Pipeline. arXiv preprint arXiv:2101.08751 (2021)."},{"key":"e_1_3_2_1_12_1","unstructured":"Pengcheng He Xiaodong Liu Jianfeng Gao and Weizhu Chen. 2020. DeBERTa: Decoding-enhanced BERT with Disentangled Attention. arXiv:2006.03654 [cs.CL]  Pengcheng He Xiaodong Liu Jianfeng Gao and Weizhu Chen. 2020. DeBERTa: Decoding-enhanced BERT with Disentangled Attention. arXiv:2006.03654 [cs.CL]"},{"key":"e_1_3_2_1_13_1","volume-title":"Overview of COLIEE 2017. In COLIEE 2017 (EPiC Series in Computing","volume":"8","author":"Kano Yoshinobu","unstructured":"Yoshinobu Kano , M. Kim , R. Goebel , and K. Satoh . 2017 . Overview of COLIEE 2017. In COLIEE 2017 (EPiC Series in Computing , vol. 47). 1-- 8 . Yoshinobu Kano, M. Kim, R. Goebel, and K. Satoh. 2017. Overview of COLIEE 2017. In COLIEE 2017 (EPiC Series in Computing, vol. 47). 1--8."},{"key":"e_1_3_2_1_14_1","volume-title":"JSAI International Symposium on Artificial Intelligence. 177--192","author":"Kano Yoshinobu","year":"2018","unstructured":"Yoshinobu Kano , Mi-Young Kim , Masaharu Yoshioka , Yao Lu , Juliano Rabelo , Naoki Kiyota , Randy Goebel , and Ken Satoh . 2018 . COLIEE-2018: Evaluation of the competition on legal information extraction and entailment . In JSAI International Symposium on Artificial Intelligence. 177--192 . Yoshinobu Kano, Mi-Young Kim, Masaharu Yoshioka, Yao Lu, Juliano Rabelo, Naoki Kiyota, Randy Goebel, and Ken Satoh. 2018. COLIEE-2018: Evaluation of the competition on legal information extraction and entailment. In JSAI International Symposium on Artificial Intelligence. 177--192."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.171"},{"key":"e_1_3_2_1_16_1","volume-title":"A Benchmark for Lease Contract Review. arXiv preprint arXiv:2010.10386","author":"Leivaditi Spyretta","year":"2020","unstructured":"Spyretta Leivaditi , Julien Rossi , and Evangelos Kanoulas . 2020. A Benchmark for Lease Contract Review. arXiv preprint arXiv:2010.10386 ( 2020 ). Spyretta Leivaditi, Julien Rossi, and Evangelos Kanoulas. 2020. A Benchmark for Lease Contract Review. arXiv preprint arXiv:2010.10386 (2020)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_2_1_18_1","volume-title":"Pyserini: An Easy-to-Use Python Toolkit to Support Replicable IR Research with Sparse and Dense Representations. arXiv preprint arXiv:2102.10073","author":"Lin Jimmy","year":"2021","unstructured":"Jimmy Lin , Xueguang Ma , Sheng-Chieh Lin , Jheng-Hong Yang , Ronak Pradeep , and Rodrigo Nogueira . 2021 . Pyserini: An Easy-to-Use Python Toolkit to Support Replicable IR Research with Sparse and Dense Representations. arXiv preprint arXiv:2102.10073 (2021). Jimmy Lin, Xueguang Ma, Sheng-Chieh Lin, Jheng-Hong Yang, Ronak Pradeep, and Rodrigo Nogueira. 2021. Pyserini: An Easy-to-Use Python Toolkit to Support Replicable IR Research with Sparse and Dense Representations. arXiv preprint arXiv:2102.10073 (2021)."},{"key":"e_1_3_2_1_19_1","volume-title":"Pretrained transformers for text ranking: Bert and beyond. arXiv preprint arXiv:2010.06467","author":"Lin Jimmy","year":"2020","unstructured":"Jimmy Lin , Rodrigo Nogueira , and Andrew Yates . 2020. Pretrained transformers for text ranking: Bert and beyond. arXiv preprint arXiv:2010.06467 ( 2020 ). Jimmy Lin, Rodrigo Nogueira, and Andrew Yates. 2020. Pretrained transformers for text ranking: Bert and beyond. arXiv preprint arXiv:2010.06467 (2020)."},{"key":"e_1_3_2_1_20_1","volume-title":"Learning from Very Few Samples: A Survey. arXiv preprint arXiv:2009.02653","author":"Lu Jiang","year":"2020","unstructured":"Jiang Lu , Pinghua Gong , Jieping Ye , and Changshui Zhang . 2020. Learning from Very Few Samples: A Survey. arXiv preprint arXiv:2009.02653 ( 2020 ). Jiang Lu, Pinghua Gong, Jieping Ye, and Changshui Zhang. 2020. Learning from Very Few Samples: A Survey. arXiv preprint arXiv:2009.02653 (2020)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3437963.3441777"},{"key":"e_1_3_2_1_22_1","volume-title":"Distributed representations of words and phrases and their compositionality. arXiv preprint arXiv:1310.4546","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov . 2013. Distributed representations of words and phrases and their compositionality. arXiv preprint arXiv:1310.4546 ( 2013 ). Tomas Mikolov. 2013. Distributed representations of words and phrases and their compositionality. arXiv preprint arXiv:1310.4546 (2013)."},{"key":"e_1_3_2_1_23_1","volume-title":"Efficient Estimation of Word Representations in Vector Space. arXiv preprint arXiv:1301.3781","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov , Kai Chen , Greg Corrado , and Jeffrey Dean . 2013. Efficient Estimation of Word Representations in Vector Space. arXiv preprint arXiv:1301.3781 ( 2013 ). Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient Estimation of Word Representations in Vector Space. arXiv preprint arXiv:1301.3781 (2013)."},{"key":"e_1_3_2_1_24_1","volume-title":"JNLP Team: Deep Learning for Legal Processing in COLIEE 2020","author":"Nguyen Ha-Thanh","year":"2020","unstructured":"Ha-Thanh Nguyen , Hai-Yen Thi Vuong , Phuong Minh Nguyen , Binh Tran Dang , Quan Minh Bui , Sinh Trong Vu , Chau Minh Nguyen , Vu Tran , Ken Satoh , and Minh Le Nguyen . 2020 . JNLP Team: Deep Learning for Legal Processing in COLIEE 2020 . arXiv preprint arXiv:2011.08071 (2020). Ha-Thanh Nguyen, Hai-Yen Thi Vuong, Phuong Minh Nguyen, Binh Tran Dang, Quan Minh Bui, Sinh Trong Vu, Chau Minh Nguyen, Vu Tran, Ken Satoh, and Minh Le Nguyen. 2020. JNLP Team: Deep Learning for Legal Processing in COLIEE 2020. arXiv preprint arXiv:2011.08071 (2020)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.63"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W19-4302"},{"key":"e_1_3_2_1_27_1","volume-title":"TREC 2020: When all you got is a hammer... Deep Learning, Health Misinformation, and Precision Medicine. Corpus 5, d3 ([n. d.]), d2.","author":"Pradeep Ronak","unstructured":"Ronak Pradeep , Xueguang Ma , Xinyu Zhang , Hang Cui , Ruizhou Xu , Rodrigo Nogueira , and Jimmy Lin . [n.d.]. H2oloo at TREC 2020: When all you got is a hammer... Deep Learning, Health Misinformation, and Precision Medicine. Corpus 5, d3 ([n. d.]), d2. Ronak Pradeep, Xueguang Ma, Xinyu Zhang, Hang Cui, Ruizhou Xu, Rodrigo Nogueira, and Jimmy Lin. [n.d.]. H2oloo at TREC 2020: When all you got is a hammer... Deep Learning, Health Misinformation, and Precision Medicine. Corpus 5, d3 ([n. d.]), d2."},{"key":"e_1_3_2_1_28_1","volume-title":"International Workshop on Juris-informatics (JURISIN) associated with JSAI International Symposia on AI (JSAI-isAI)","author":"Rabelo J.","year":"2020","unstructured":"J. Rabelo , M.Y. Kim , and R. Goebel . 2020. Application of text entailment techniques in COLIEE 2020 . International Workshop on Juris-informatics (JURISIN) associated with JSAI International Symposia on AI (JSAI-isAI) ( 2020 ). J. Rabelo, M.Y. Kim, and R. Goebel. 2020. Application of text entailment techniques in COLIEE 2020. International Workshop on Juris-informatics (JURISIN) associated with JSAI International Symposia on AI (JSAI-isAI) (2020)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3322640.3326741"},{"key":"e_1_3_2_1_30_1","volume-title":"JSAI International Symposium on Artificial Intelligence. 34--49","author":"Rabelo Juliano","year":"2019","unstructured":"Juliano Rabelo , Mi-Young Kim , Randy Goebel , Masaharu Yoshioka , Yoshinobu Kano , and Ken Satoh . 2019 . A Summary of the COLIEE 2019 Competition . In JSAI International Symposium on Artificial Intelligence. 34--49 . Juliano Rabelo, Mi-Young Kim, Randy Goebel, Masaharu Yoshioka, Yoshinobu Kano, and Ken Satoh. 2019. A Summary of the COLIEE 2019 Competition. In JSAI International Symposium on Artificial Intelligence. 34--49."},{"key":"e_1_3_2_1_31_1","volume-title":"COLIEE 2020: Methods for Legal Document Retrieval and Entailment.","author":"Rabelo Juliano","year":"2020","unstructured":"Juliano Rabelo , Mi-Young Kim , Randy Goebel , Masaharu Yoshioka , Yoshinobu Kano , and Ken Satoh . 2020 . COLIEE 2020: Methods for Legal Document Retrieval and Entailment. (2020). Juliano Rabelo, Mi-Young Kim, Randy Goebel, Masaharu Yoshioka, Yoshinobu Kano, and Ken Satoh. 2020. COLIEE 2020: Methods for Legal Document Retrieval and Entailment. (2020)."},{"key":"e_1_3_2_1_32_1","volume-title":"Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford , Jong Wook Kim , Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021 . Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020 (2021). Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020 (2021)."},{"key":"e_1_3_2_1_33_1","first-page":"1","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel , Noam Shazeer , Adam Roberts , Katherine Lee , Sharan Narang , Michael Matena , Yanqi Zhou , Wei Li , and Peter J. Liu . 2020 . Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer . Journal of Machine Learning Research 21 , 140 (2020), 1 -- 67 . http:\/\/jmlr.org\/papers\/v21\/20-074.html Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research 21, 140 (2020), 1--67. http:\/\/jmlr.org\/papers\/v21\/20-074.html","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_34_1","volume-title":"Overview of the TREC 2019 Precision Medicine Track. The ... text REtrieval conference : TREC. Text REtrieval Conference 26","author":"Roberts Kirk","year":"2019","unstructured":"Kirk Roberts , Dina Demner-Fushman , E. Voorhees , W. Hersh , Steven Bedrick , Alexander J. Lazar , and S. Pant . 2019 . Overview of the TREC 2019 Precision Medicine Track. The ... text REtrieval conference : TREC. Text REtrieval Conference 26 ( 2019 ). Kirk Roberts, Dina Demner-Fushman, E. Voorhees, W. Hersh, Steven Bedrick, Alexander J. Lazar, and S. Pant. 2019. Overview of the TREC 2019 Precision Medicine Track. The ... text REtrieval conference : TREC. Text REtrieval Conference 26 (2019)."},{"key":"e_1_3_2_1_35_1","first-page":"109","article-title":"Okapi at TREC-3","volume":"109","author":"Robertson Stephen E","year":"1995","unstructured":"Stephen E Robertson , Steve Walker , Susan Jones , Micheline M Hancock-Beaulieu , Mike Gatford , 1995 . Okapi at TREC-3 . Nist Special Publication Sp 109 (1995), 109 . Stephen E Robertson, Steve Walker, Susan Jones, Micheline M Hancock-Beaulieu, Mike Gatford, et al. 1995. Okapi at TREC-3. Nist Special Publication Sp 109 (1995), 109.","journal-title":"Nist Special Publication Sp"},{"key":"e_1_3_2_1_36_1","volume-title":"Exploiting cloze questions for few-shot text classification and natural language inference. arXiv preprint arXiv:2001.07676","author":"Schick Timo","year":"2020","unstructured":"Timo Schick and Hinrich Sch\u00fctze . 2020. Exploiting cloze questions for few-shot text classification and natural language inference. arXiv preprint arXiv:2001.07676 ( 2020 ). Timo Schick and Hinrich Sch\u00fctze. 2020. Exploiting cloze questions for few-shot text classification and natural language inference. arXiv preprint arXiv:2001.07676 (2020)."},{"key":"e_1_3_2_1_37_1","volume-title":"Customizing Contextualized Language Models for Legal Document Reviews. In 2020 IEEE International Conference on Big Data (Big Data). IEEE, 2139--2148","author":"Shaghaghian Shohreh","year":"2020","unstructured":"Shohreh Shaghaghian , Luna Yue Feng , Borna Jafarpour , and Nicolai Pogrebnyakov . 2020 . Customizing Contextualized Language Models for Legal Document Reviews. In 2020 IEEE International Conference on Big Data (Big Data). IEEE, 2139--2148 . Shohreh Shaghaghian, Luna Yue Feng, Borna Jafarpour, and Nicolai Pogrebnyakov. 2020. Customizing Contextualized Language Models for Legal Document Reviews. In 2020 IEEE International Conference on Big Data (Big Data). IEEE, 2139--2148."},{"key":"e_1_3_2_1_38_1","volume-title":"Improving and Simplifying Pattern Exploiting Training. arXiv preprint arXiv:2103.11955","author":"Tam Derek","year":"2021","unstructured":"Derek Tam , Rakesh R Menon , Mohit Bansal , Shashank Srivastava , and Colin Raffel . 2021. Improving and Simplifying Pattern Exploiting Training. arXiv preprint arXiv:2103.11955 ( 2021 ). Derek Tam, Rakesh R Menon, Mohit Bansal, Shashank Srivastava, and Colin Raffel. 2021. Improving and Simplifying Pattern Exploiting Training. arXiv preprint arXiv:2103.11955 (2021)."},{"key":"e_1_3_2_1_39_1","volume-title":"BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models. arXiv preprint arXiv:2104.08663 (4","author":"Thakur Nandan","year":"2021","unstructured":"Nandan Thakur , Nils Reimers , Andreas R\u00fcckl\u00e9 , Abhishek Srivastava , and Iryna Gurevych . 2021 . BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models. arXiv preprint arXiv:2104.08663 (4 2021). https:\/\/arxiv.org\/abs\/2104.08663 Nandan Thakur, Nils Reimers, Andreas R\u00fcckl\u00e9, Abhishek Srivastava, and Iryna Gurevych. 2021. BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models. arXiv preprint arXiv:2104.08663 (4 2021). https:\/\/arxiv.org\/abs\/2104.08663"},{"key":"e_1_3_2_1_40_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention is All you Need. In NIPS.  Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention is All you Need. In NIPS."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"crossref","unstructured":"Giorgos Vernikos Katerina Margatina Alexandra Chronopoulou and Ion Androutsopoulos. 2020. Domain Adversarial Fine-Tuning as an Effective Regularizer. arXiv:2009.13366 [cs.LG]  Giorgos Vernikos Katerina Margatina Alexandra Chronopoulou and Ion Androutsopoulos. 2020. Domain Adversarial Fine-Tuning as an Effective Regularizer. arXiv:2009.13366 [cs.LG]","DOI":"10.18653\/v1\/2020.findings-emnlp.278"},{"key":"e_1_3_2_1_42_1","volume-title":"Proceedings of the Thirteenth Text REtrieval Conference, TREC 2004","author":"Voorhees Ellen M.","year":"2004","unstructured":"Ellen M. Voorhees . 2004 . Overview of the TREC 2004 Robust Track . Proceedings of the Thirteenth Text REtrieval Conference, TREC 2004 , Gaithersburg, Maryland, November 16--19 , 2004 (2004). Ellen M. Voorhees. 2004. Overview of the TREC 2004 Robust Track. Proceedings of the Thirteenth Text REtrieval Conference, TREC 2004, Gaithersburg, Maryland, November 16--19, 2004 (2004)."},{"key":"e_1_3_2_1_43_1","volume-title":"CAIL2018: A Large-Scale Legal Dataset for Judgment Prediction. arXiv:1807","author":"Xiao Chaojun","year":"2018","unstructured":"Chaojun Xiao , Haoxi Zhong , Zhipeng Guo , Cunchao Tu , Zhiyuan Liu , Maosong Sun , Yansong Feng , Xianpei Han , Zhen Hu , Heng Wang , and Jianfeng Xu . 2018 . CAIL2018: A Large-Scale Legal Dataset for Judgment Prediction. arXiv:1807 .02478 (2018). Chaojun Xiao, Haoxi Zhong, Zhipeng Guo, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Yansong Feng, Xianpei Han, Zhen Hu, Heng Wang, and Jianfeng Xu. 2018. CAIL2018: A Large-Scale Legal Dataset for Judgment Prediction. arXiv:1807.02478 (2018)."},{"key":"e_1_3_2_1_44_1","volume-title":"Effects of inserting domain vocabulary and fine-tuning BERT for German legal language. Master's thesis","author":"Yeung Chin Man","unstructured":"Chin Man Yeung . 2019. Effects of inserting domain vocabulary and fine-tuning BERT for German legal language. Master's thesis . University of Twente . Chin Man Yeung. 2019. Effects of inserting domain vocabulary and fine-tuning BERT for German legal language. Master's thesis. University of Twente."},{"key":"e_1_3_2_1_45_1","unstructured":"Wenpeng Yin Jamaal Hay and Dan Roth. 2019. Benchmarking Zero-shot Text Classification: Datasets Evaluation and Entailment Approach. arXiv:1909.00161 [cs.CL]  Wenpeng Yin Jamaal Hay and Dan Roth. 2019. Benchmarking Zero-shot Text Classification: Datasets Evaluation and Entailment Approach. arXiv:1909.00161 [cs.CL]"},{"key":"e_1_3_2_1_46_1","volume-title":"Dragomir Radev, Richard Socher, and Caiming Xiong.","author":"Yin Wenpeng","year":"2020","unstructured":"Wenpeng Yin , Nazneen Fatema Rajani , Dragomir Radev, Richard Socher, and Caiming Xiong. 2020 . Universal Natural Language Processing with Limited Annotations : Try Few-shot Textual Entailment as a Start . arXiv:2010.02584 [cs.CL] Wenpeng Yin, Nazneen Fatema Rajani, Dragomir Radev, Richard Socher, and Caiming Xiong. 2020. Universal Natural Language Processing with Limited Annotations: Try Few-shot Textual Entailment as a Start. arXiv:2010.02584 [cs.CL]"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00097"},{"key":"e_1_3_2_1_48_1","volume-title":"Proceedings of the 1st Workshop on NLP for COVID-19 at ACL","author":"Zhang Edwin","year":"2020","unstructured":"Edwin Zhang , Nikhil Gupta , Rodrigo Nogueira , Kyunghyun Cho , and Jimmy Lin . 2020 . Rapidly Deploying a Neural Search Engine for the COVID-19 Open Research Dataset . In Proceedings of the 1st Workshop on NLP for COVID-19 at ACL 2020. Edwin Zhang, Nikhil Gupta, Rodrigo Nogueira, Kyunghyun Cho, and Jimmy Lin. 2020. Rapidly Deploying a Neural Search Engine for the COVID-19 Open Research Dataset. In Proceedings of the 1st Workshop on NLP for COVID-19 at ACL 2020."},{"key":"e_1_3_2_1_49_1","unstructured":"Tianyi Zhang Felix Wu Arzoo Katiyar Kilian Q. Weinberger and Yoav Artzi. 2021. Revisiting Few-sample BERT Fine-tuning. arXiv:2006.05987 [cs.CL]  Tianyi Zhang Felix Wu Arzoo Katiyar Kilian Q. Weinberger and Yoav Artzi. 2021. Revisiting Few-sample BERT Fine-tuning. arXiv:2006.05987 [cs.CL]"},{"key":"e_1_3_2_1_50_1","volume-title":"How Does NLP Benefit Legal System: A Summary of Legal Artificial Intelligence. arXiv:2004.12158","author":"Zhong Haoxi","year":"2020","unstructured":"Haoxi Zhong , Chaojun Xiao , Cunchao Tu , Tianyang Zhang , Zhiyuan Liu , and Maosong Sun . 2020. How Does NLP Benefit Legal System: A Summary of Legal Artificial Intelligence. arXiv:2004.12158 ( 2020 ). Haoxi Zhong, Chaojun Xiao, Cunchao Tu, Tianyang Zhang, Zhiyuan Liu, and Maosong Sun. 2020. How Does NLP Benefit Legal System: A Summary of Legal Artificial Intelligence. arXiv:2004.12158 (2020)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6519"}],"event":{"name":"ICAIL '21: Eighteenth International Conference for Artificial Intelligence and Law","location":"S\u00e3o Paulo Brazil","acronym":"ICAIL '21","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence"]},"container-title":["Proceedings of the Eighteenth International Conference on Artificial Intelligence and Law"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3462757.3466103","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3462757.3466103","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T21:28:31Z","timestamp":1750195711000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3462757.3466103"}},"subtitle":["zero-shot models for legal case entailment"],"short-title":[],"issued":{"date-parts":[[2021,6,21]]},"references-count":51,"alternative-id":["10.1145\/3462757.3466103","10.1145\/3462757"],"URL":"https:\/\/doi.org\/10.1145\/3462757.3466103","relation":{},"subject":[],"published":{"date-parts":[[2021,6,21]]},"assertion":[{"value":"2021-07-27","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}