{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,7]],"date-time":"2024-09-07T20:49:24Z","timestamp":1725742164876},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,12,14]],"date-time":"2023-12-14T00:00:00Z","timestamp":1702512000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,12,14]],"date-time":"2023-12-14T00:00:00Z","timestamp":1702512000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,12,14]]},"DOI":"10.1109\/bcd57833.2023.10466348","type":"proceedings-article","created":{"date-parts":[[2024,3,19]],"date-time":"2024-03-19T18:11:47Z","timestamp":1710871907000},"page":"201-207","source":"Crossref","is-referenced-by-count":0,"title":["Leveraging Unannotated Data to Improve Zero-Shot Question Answering in the French Legal Domain"],"prefix":"10.1109","author":[{"given":"Ahmed","family":"Touila","sequence":"first","affiliation":[{"name":"Hyperlex\/Dilitrust,Applied Research Team,Paris,France"}]},{"given":"Elie","family":"Louis","sequence":"additional","affiliation":[{"name":"Hyperlex\/Dilitrust,Applied Research Team,Paris,France"}]},{"given":"Hamza","family":"Gharbi","sequence":"additional","affiliation":[{"name":"Hyperlex\/Dilitrust,Applied Research Team,Paris,France"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.468"},{"article-title":"ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning","year":"2022","author":"Aribandi","key":"ref2"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1017\/S1351324921000395"},{"article-title":"Language Models are Few-Shot Learners","year":"2020","author":"Brown","key":"ref4"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.261"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1171"},{"key":"ref7","first-page":"2924","article-title":"BoolQ: Exploring the surprising difficulty of natural yes\/no questions","author":"Clark","year":"2019"},{"article-title":"Multi-step Retriever-Reader Interaction for Scalable Open-domain Question Answering","year":"2019","author":"Das","key":"ref8"},{"article-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","year":"2019","author":"Devlin","key":"ref9"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3039353"},{"article-title":"CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review","year":"2021","author":"Hendrycks","key":"ref11"},{"article-title":"spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing","year":"2017","author":"Honnibal","key":"ref12"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/762"},{"journal-title":"French-SQuAD : French Machine Reading for Question Answering","year":"2021","author":"Kabbadj","key":"ref14"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"article-title":"Project PIAF: Building a Native French Question-Answering Dataset","year":"2020","author":"Keraron","key":"ref16"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00276"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1549"},{"article-title":"ALBERT: A Lite BERT for Self-supervised Learning of Language Representations","year":"2020","author":"Lan","key":"ref20"},{"journal-title":"FlauBERT: Unsupervised Language Model Pre-training for French. page 12","author":"Le","key":"ref21"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.297"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1612"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1484"},{"article-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach","year":"2019","author":"Liu","key":"ref25"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.645"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/1165485.1165536"},{"article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","year":"2020","author":"Raffel","key":"ref28"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1264"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1561\/1500000019"},{"article-title":"Multitask Prompted Training Enables Zero-Shot Task Generalization","year":"2022","author":"Sanh","key":"ref32"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1177\/107769905303000401"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1145\/3462757.3466102"},{"article-title":"Yuan 1.0: LargeScale Pre-trained Language Model in Zero-Shot and Few-Shot Learning","year":"2021","author":"Wu","key":"ref35"}],"event":{"name":"2023 IEEE\/ACIS 8th International Conference on Big Data, Cloud Computing, and Data Science (BCD)","start":{"date-parts":[[2023,12,14]]},"location":"Hochimin City, Vietnam","end":{"date-parts":[[2023,12,16]]}},"container-title":["2023 IEEE\/ACIS 8th International Conference on Big Data, Cloud Computing, and Data Science (BCD)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10466269\/10466270\/10466348.pdf?arnumber=10466348","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,26]],"date-time":"2024-03-26T14:17:59Z","timestamp":1711462679000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10466348\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,14]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/bcd57833.2023.10466348","relation":{},"subject":[],"published":{"date-parts":[[2023,12,14]]}}}