{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T15:39:20Z","timestamp":1766158760914,"version":"3.40.3"},"publisher-location":"Cham","reference-count":17,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030884826"},{"type":"electronic","value":"9783030884833"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-88483-3_31","type":"book-chapter","created":{"date-parts":[[2021,10,10]],"date-time":"2021-10-10T13:55:25Z","timestamp":1633874125000},"page":"388-399","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Enhanced Few-Shot Learning with Multiple-Pattern-Exploiting Training"],"prefix":"10.1007","author":[{"given":"Jiali","family":"Zeng","sequence":"first","affiliation":[]},{"given":"Yufan","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Shuangzhi","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Mu","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,10,6]]},"reference":[{"key":"31_CR1","doi-asserted-by":"publisher","unstructured":"Bowman, S.R., Angeli, G., Potts, C., Manning, C.D.: A large annotated corpus for learning natural language inference. In: M\u00e0rquez, L., Callison-Burch, C., Su, J., Pighin, D., Marton, Y. (eds.) Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, 17\u201321 September 2015, Lisbon, Portugal, pp. 632\u2013642. The Association for Computational Linguistics (2015). https:\/\/doi.org\/10.18653\/v1\/d15-1075","DOI":"10.18653\/v1\/d15-1075"},{"key":"31_CR2","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"31_CR3","doi-asserted-by":"crossref","unstructured":"Gururangan, S., et al.: Don\u2019t stop pretraining: adapt language models to domains and tasks. arXiv preprint arXiv:2004.10964 (2020)","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"31_CR4","doi-asserted-by":"publisher","unstructured":"Hu, H., Richardson, K., Xu, L., Li, L., K\u00fcbler, S., Moss, L.S.: OCNLI: original Chinese natural language inference. In: Cohn, T., He, Y., Liu, Y. (eds.) Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, EMNLP 2020, Online Event, 16\u201320 November 2020. Findings of ACL, vol. EMNLP 2020, pp. 3512\u20133526. Association for Computational Linguistics (2020). https:\/\/doi.org\/10.18653\/v1\/2020.findings-emnlp.314","DOI":"10.18653\/v1\/2020.findings-emnlp.314"},{"key":"31_CR5","doi-asserted-by":"publisher","unstructured":"Lai, G., Xie, Q., Liu, H., Yang, Y., Hovy, E.H.: RACE: large-scale reading comprehension dataset from examinations. In: Palmer, M., Hwa, R., Riedel, S. (eds.) Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP 2017, 9\u201311 September 2017, Copenhagen, Denmark, pp. 785\u2013794. Association for Computational Linguistics (2017). https:\/\/doi.org\/10.18653\/v1\/d17-1082","DOI":"10.18653\/v1\/d17-1082"},{"key":"31_CR6","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: ALBERT: a lite BERT for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019)"},{"key":"31_CR7","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"31_CR8","unstructured":"Liu, Y., et al.: RoBERTa: A robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"31_CR9","doi-asserted-by":"publisher","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100, 000+ Questions for machine comprehension of text. In: Su, J., Carreras, X., Duh, K. (eds.) Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, 1\u20134 November 2016, Austin, Texas, USA, pp. 2383\u20132392. The Association for Computational Linguistics (2016). https:\/\/doi.org\/10.18653\/v1\/d16-1264","DOI":"10.18653\/v1\/d16-1264"},{"key":"31_CR10","doi-asserted-by":"crossref","unstructured":"Schick, T., Sch\u00fctze, H.: Exploiting cloze-questions for few-shot text classification and natural language inference. In: Merlo, P., Tiedemann, J., Tsarfaty, R. (eds.) Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL 2021, Online, 19\u201323 April 2021, pp. 255\u2013269. Association for Computational Linguistics (2021). https:\/\/www.aclweb.org\/anthology\/2021.eacl-main.20\/","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"31_CR11","doi-asserted-by":"crossref","unstructured":"Tam, D., Menon, R.R., Bansal, M., Srivastava, S., Raffel, C.: Improving and simplifying pattern exploiting training. CoRR abs\/2103.11955 (2021). https:\/\/arxiv.org\/abs\/2103.11955","DOI":"10.18653\/v1\/2021.emnlp-main.407"},{"key":"31_CR12","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S.R.: GLUE: a multi-task benchmark and analysis platform for natural language understanding. In: 7th International Conference on Learning Representations, ICLR 2019, New Orleans, 6\u20139 May 2019, LA, USA. OpenReview.net (2019). https:\/\/openreview.net\/forum?id=rJ4km2R5t7"},{"key":"31_CR13","unstructured":"Whang, T., Lee, D., Lee, C., Yang, K., Oh, D., Lim, H.: Domain adaptive training BERT for response selection. arXiv preprint arXiv:1908.04812 (2019)"},{"key":"31_CR14","doi-asserted-by":"publisher","unstructured":"Xu, L., et al.: CLUE: a Chinese language understanding evaluation benchmark. In: Scott, D., Bel, N., Zong, C. (eds.) Proceedings of the 28th International Conference on Computational Linguistics, COLING 2020, 8\u201313 December 2020, Barcelona, Spain (Online), pp. 4762\u20134772. International Committee on Computational Linguistics (2020). https:\/\/doi.org\/10.18653\/v1\/2020.coling-main.419","DOI":"10.18653\/v1\/2020.coling-main.419"},{"key":"31_CR15","unstructured":"Xu, L., et al.: FewCLUE: a Chinese few-shot learning evaluation benchmark. CoRR abs\/2107.07498 (2021). arXiv: 2107.07498"},{"key":"31_CR16","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"31_CR17","doi-asserted-by":"publisher","unstructured":"Zheng, C., Huang, M., Sun, A.: ChiD: a large-scale Chinese idiom dataset for cloze test. In: Korhonen, A., Traum, D.R., M\u00e0rquez, L. (eds.) Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, 28 July\u20132 August 2019, Florence, Italy, vol. 1, Long Papers, pp. 778\u2013787. Association for Computational Linguistics (2019). https:\/\/doi.org\/10.18653\/v1\/p19-1075","DOI":"10.18653\/v1\/p19-1075"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Chinese Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-88483-3_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T15:35:33Z","timestamp":1710257733000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-88483-3_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030884826","9783030884833"],"references-count":17,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-88483-3_31","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"6 October 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLPCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF International Conference on Natural Language Processing and Chinese Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Qingdao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 October 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nlpcc2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/tcci.ccf.org.cn\/conference\/2021\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Softconf","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"446","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"66","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"15% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1.5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"23 poster papers and 27 workshop papers are also included.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}