{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T16:26:55Z","timestamp":1743006415565,"version":"3.40.3"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031446955"},{"type":"electronic","value":"9783031446962"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-44696-2_4","type":"book-chapter","created":{"date-parts":[[2023,10,7]],"date-time":"2023-10-07T09:03:59Z","timestamp":1696669439000},"page":"43-54","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Prompt-Free Few-Shot Learning with\u00a0ELECTRA for\u00a0Acceptability Judgment"],"prefix":"10.1007","author":[{"given":"Linqin","family":"Li","sequence":"first","affiliation":[]},{"given":"Zicheng","family":"Li","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Shoushan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Guodong","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,8]]},"reference":[{"key":"4_CR1","unstructured":"Brown, T., et al.: Language models are few-shot learners. In: Advances in Neural Information Processing Systems, vol. 33, pp. 1877\u20131901 (2020)"},{"key":"4_CR2","unstructured":"Clark, K., Luong, M.T., Le, Q.V., Manning, C.D.: ELECTRA: pre-training text encoders as discriminators rather than generators. In: International Conference on Learning Representations (2020)"},{"key":"4_CR3","unstructured":"Dahlmeier, D., Ng, H.T.: Correcting semantic collocation errors with L1-induced paraphrases. In: Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, pp. 107\u2013117 (2011)"},{"key":"4_CR4","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171\u20134186 (2019)"},{"key":"4_CR5","unstructured":"Dodge, J., Ilharco, G., Schwartz, R., Farhadi, A., Hajishirzi, H., Smith, N.: Fine-tuning pretrained language models: weight initializations, data orders, and early stopping. arXiv preprint arXiv:2002.06305 (2020)"},{"key":"4_CR6","doi-asserted-by":"crossref","unstructured":"Gao, T., Fisch, A., Chen, D.: Making pre-trained language models better few-shot learners. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 3816\u20133830 (2021)","DOI":"10.18653\/v1\/2021.acl-long.295"},{"key":"4_CR7","doi-asserted-by":"crossref","unstructured":"Hambardzumyan, K., Khachatrian, H., May, J.: WARP: word-level adversarial reprogramming. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 4921\u20134933 (2021)","DOI":"10.18653\/v1\/2021.acl-long.381"},{"key":"4_CR8","doi-asserted-by":"crossref","unstructured":"Kann, K., Rothe, S., Filippova, K.: Sentence-level fluency evaluation: references help, but can be spared! In: Proceedings of the 22nd Conference on Computational Natural Language Learning, pp. 313\u2013323 (2018)","DOI":"10.18653\/v1\/K18-1031"},{"key":"4_CR9","doi-asserted-by":"crossref","unstructured":"Karimi Mahabadi, R., et al.: Prompt-free and efficient few-shot learning with language models. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 3638\u20133652 (2022)","DOI":"10.18653\/v1\/2022.acl-long.254"},{"key":"4_CR10","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: ALBERT: a lite BERT for self-supervised learning of language representations. In: International Conference on Learning Representations (2020)"},{"key":"4_CR11","unstructured":"Li, Z., Li, S., Zhou, G.: Pre-trained token-replaced detection model as few-shot learner. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 3274\u20133284 (2022)"},{"issue":"9","key":"4_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023)","journal-title":"ACM Comput. Surv."},{"key":"4_CR13","unstructured":"Liu, X., et al.: GPT understands, too. arXiv preprint arXiv:2103.10385 (2021)"},{"key":"4_CR14","unstructured":"Liu, Y., et al.: ROBERTA: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"issue":"2","key":"4_CR15","doi-asserted-by":"publisher","first-page":"442","DOI":"10.1016\/0005-2795(75)90109-9","volume":"405","author":"B Matthews","year":"1975","unstructured":"Matthews, B.: Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochimica et Biophysica Acta (BBA) - Protein Struct. 405(2), 442\u2013451 (1975)","journal-title":"Biochimica et Biophysica Acta (BBA) - Protein Struct."},{"key":"4_CR16","doi-asserted-by":"crossref","unstructured":"Min, S., et al.: Rethinking the role of demonstrations: what makes in-context learning work? arXiv preprint arXiv:2202.12837 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.759"},{"key":"4_CR17","doi-asserted-by":"crossref","unstructured":"Schick, T., Sch\u00fctze, H.: Exploiting cloze-questions for few-shot text classification and natural language inference. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp. 255\u2013269 (2021)","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"4_CR18","doi-asserted-by":"crossref","unstructured":"Schick, T., Sch\u00fctze, H.: It\u2019s not just size that matters: small language models are also few-shot learners. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2339\u20132352 (2021)","DOI":"10.18653\/v1\/2021.naacl-main.185"},{"key":"4_CR19","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1007\/s11633-022-1331-6","volume":"19","author":"T Sun","year":"2022","unstructured":"Sun, T., Liu, X., Qiu, X., Huang, X.: Paradigm shift in natural language processing. Mach. Intell. Res. 19, 169\u2013183 (2022)","journal-title":"Mach. Intell. Res."},{"key":"4_CR20","doi-asserted-by":"crossref","unstructured":"Trotta, D., Guarasci, R., Leonardelli, E., Tonelli, S.: Monolingual and cross-lingual acceptability judgments with the Italian CoLA corpus. In: Findings of the Association for Computational Linguistics, EMNLP 2021, pp. 2929\u20132940 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.250"},{"key":"4_CR21","unstructured":"Wang, W., et al.: StructBERT: incorporating language structures into pre-training for deep language understanding. In: International Conference on Learning Representations (2020)"},{"key":"4_CR22","doi-asserted-by":"publisher","first-page":"625","DOI":"10.1162\/tacl_a_00290","volume":"7","author":"A Warstadt","year":"2019","unstructured":"Warstadt, A., Singh, A., Bowman, S.: Neural network acceptability judgments. Trans. Assoc. Comput. Linguist. 7, 625\u2013641 (2019)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Xia, M., Artetxe, M., Du, J., Chen, D., Stoyanov, V.: Prompting ELECTRA: few-shot learning with discriminative pre-trained models. arXiv preprint arXiv:2205.15223 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.780"},{"key":"4_CR24","doi-asserted-by":"crossref","unstructured":"Yao, Y., et al.: Prompt tuning for discriminative pre-trained language models. In: Findings of the Association for Computational Linguistics, ACL 2022, pp. 3468\u20133473 (2022)","DOI":"10.18653\/v1\/2022.findings-acl.273"},{"key":"4_CR25","unstructured":"Zhang, N., et al.: Differentiable prompt makes pre-trained language models better few-shot learners. In: International Conference on Learning Representations (2022)"},{"key":"4_CR26","unstructured":"Zhang, T., Wu, F., Katiyar, A., Weinberger, K.Q., Artzi, Y.: Revisiting few-sample BERT fine-tuning. In: International Conference on Learning Representations (2021)"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Chinese Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-44696-2_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,7]],"date-time":"2023-10-07T09:04:30Z","timestamp":1696669470000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-44696-2_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031446955","9783031446962"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-44696-2_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"8 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLPCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF International Conference on Natural Language Processing and Chinese Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Foshan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nlpcc2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/tcci.ccf.org.cn\/conference\/2023\/index.php","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Softconf","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"478","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"143","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"30% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}