{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T10:33:55Z","timestamp":1763202835348,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,21]],"date-time":"2023-10-21T00:00:00Z","timestamp":1697846400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,21]]},"DOI":"10.1145\/3583780.3614742","type":"proceedings-article","created":{"date-parts":[[2023,10,21]],"date-time":"2023-10-21T07:45:26Z","timestamp":1697874326000},"page":"5111-5116","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["HugNLP: A Unified and Comprehensive Library for Natural Language Processing"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6006-053X","authenticated-orcid":false,"given":"Jianing","family":"Wang","sequence":"first","affiliation":[{"name":"East China Normal University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6563-1215","authenticated-orcid":false,"given":"Nuo","family":"Chen","sequence":"additional","affiliation":[{"name":"East China Normal University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5207-818X","authenticated-orcid":false,"given":"Qiushi","family":"Sun","sequence":"additional","affiliation":[{"name":"National University of Singapore, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6443-2246","authenticated-orcid":false,"given":"Wenkang","family":"Huang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1010-9678","authenticated-orcid":false,"given":"Chengyu","family":"Wang","sequence":"additional","affiliation":[{"name":"Alibaba Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5603-2680","authenticated-orcid":false,"given":"Ming","family":"Gao","sequence":"additional","affiliation":[{"name":"East China Normal University, Shanghai, China"}]}],"member":"320","published-online":{"date-parts":[[2023,10,21]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"A Survey. CoRR","author":"Amini Massih-Reza","year":"2040","unstructured":"Massih-Reza Amini , Vasilii Feofanov , Lo\"i c Pauletto, Emilie Devijver , and Yury Maximov . 2022. Self-Training : A Survey. CoRR , Vol. abs\/ 2202 .1 2040 (2022). Massih-Reza Amini, Vasilii Feofanov, Lo\"i c Pauletto, Emilie Devijver, and Yury Maximov. 2022. Self-Training: A Survey. CoRR, Vol. abs\/2202.12040 (2022)."},{"key":"e_1_3_2_1_2_1","unstructured":"Tom B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. In NeurIPS. Tom B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. In NeurIPS."},{"key":"e_1_3_2_1_3_1","first-page":"331","article-title":"Learning From Labeled And Unlabeled Data","volume":"23","author":"Chawla Nitesh V.","year":"2005","unstructured":"Nitesh V. Chawla and Grigoris I. Karakoulas . 2005 . Learning From Labeled And Unlabeled Data : An Empirical Study Across Techniques And Domains. JAIS , Vol. 23 (2005), 331 -- 366 . Nitesh V. Chawla and Grigoris I. Karakoulas. 2005. Learning From Labeled And Unlabeled Data: An Empirical Study Across Techniques And Domains. JAIS, Vol. 23 (2005), 331--366.","journal-title":"An Empirical Study Across Techniques And Domains. JAIS"},{"key":"e_1_3_2_1_4_1","unstructured":"Yiming Cui Wanxiang Che Ting Liu Bing Qin Shijin Wang and Guoping Hu. 2020. Revisiting Pre-Trained Models for Chinese Natural Language Processing. In EMNLP (Findings). 657--668. Yiming Cui Wanxiang Che Ting Liu Bing Qin Shijin Wang and Guoping Hu. 2020. Revisiting Pre-Trained Models for Chinese Natural Language Processing. In EMNLP (Findings). 657--668."},{"key":"e_1_3_2_1_5_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186.","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186."},{"key":"e_1_3_2_1_6_1","volume-title":"A Survey for In-context Learning. CoRR","author":"Dong Qingxiu","year":"2023","unstructured":"Qingxiu Dong , Lei Li , Damai Dai , Ce Zheng , Zhiyong Wu , Baobao Chang , Xu Sun , Jingjing Xu , Lei Li , and Zhifang Sui . 2023. A Survey for In-context Learning. CoRR , Vol. abs\/ 2301 .00234 ( 2023 ). https:\/\/doi.org\/10.48550\/arXiv.2301.00234 showeprint[arXiv]2301.00234 10.48550\/arXiv.2301.00234 Qingxiu Dong, Lei Li, Damai Dai, Ce Zheng, Zhiyong Wu, Baobao Chang, Xu Sun, Jingjing Xu, Lei Li, and Zhifang Sui. 2023. A Survey for In-context Learning. CoRR, Vol. abs\/2301.00234 (2023). https:\/\/doi.org\/10.48550\/arXiv.2301.00234 showeprint[arXiv]2301.00234"},{"volume-title":"CodeBERT: A Pre-Trained Model for Programming and Natural Languages","author":"Feng Zhangyin","key":"e_1_3_2_1_7_1","unstructured":"Zhangyin Feng , Daya Guo , Duyu Tang , Nan Duan , Xiaocheng Feng , Ming Gong , Linjun Shou , Bing Qin , Ting Liu , Daxin Jiang , and Ming Zhou . 2020. CodeBERT: A Pre-Trained Model for Programming and Natural Languages . In EMNLP. Association for Computational Linguistics , Online , 1536--1547. Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, and Ming Zhou. 2020. CodeBERT: A Pre-Trained Model for Programming and Natural Languages. In EMNLP. Association for Computational Linguistics, Online, 1536--1547."},{"key":"e_1_3_2_1_8_1","first-page":"1050","article-title":"Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning","volume":"48","author":"Gal Yarin","year":"2016","unstructured":"Yarin Gal and Zoubin Ghahramani . 2016 . Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning . In ICML , Vol. 48. 1050 -- 1059 . Yarin Gal and Zoubin Ghahramani. 2016. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In ICML, Vol. 48. 1050--1059.","journal-title":"ICML"},{"key":"e_1_3_2_1_9_1","unstructured":"Pengcheng He Xiaodong Liu Jianfeng Gao and Weizhu Chen. 2021. Deberta: decoding-Enhanced Bert with Disentangled Attention. In ICLR. OpenReview.net. Pengcheng He Xiaodong Liu Jianfeng Gao and Weizhu Chen. 2021. Deberta: decoding-Enhanced Bert with Disentangled Attention. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_10_1","volume-title":"ICML (Proceedings of Machine Learning Research","volume":"2799","author":"Houlsby Neil","year":"2019","unstructured":"Neil Houlsby , Andrei Giurgiu , Stanislaw Jastrzebski , Bruna Morrone , Quentin de Laroussilhe , Andrea Gesmundo , Mona Attariyan , and Sylvain Gelly . 2019 . Parameter-Efficient Transfer Learning for NLP . In ICML (Proceedings of Machine Learning Research , Vol. 97). 2790-- 2799 . Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin de Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. 2019. Parameter-Efficient Transfer Learning for NLP. In ICML (Proceedings of Machine Learning Research, Vol. 97). 2790--2799."},{"key":"e_1_3_2_1_11_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations, ICLR 2022","author":"Hu Edward J.","year":"2022","unstructured":"Edward J. Hu , Yelong Shen , Phillip Wallis , Zeyuan Allen-Zhu , Yuanzhi Li , Shean Wang , Lu Wang , and Weizhu Chen . 2022 b. LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations, ICLR 2022 , Virtual Event, April 25--29 , 2022. OpenReview.net. Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2022b. LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25--29, 2022. OpenReview.net."},{"key":"e_1_3_2_1_12_1","volume-title":"A Survey of Knowledge-Enhanced Pre-trained Language Models. CoRR","author":"Hu Linmei","year":"2022","unstructured":"Linmei Hu , Zeyi Liu , Ziwang Zhao , Lei Hou , Liqiang Nie , and Juanzi Li. 2022a. A Survey of Knowledge-Enhanced Pre-trained Language Models. CoRR , Vol. abs\/ 2211 .05994 ( 2022 ). Linmei Hu, Zeyi Liu, Ziwang Zhao, Lei Hou, Liqiang Nie, and Juanzi Li. 2022a. A Survey of Knowledge-Enhanced Pre-trained Language Models. CoRR, Vol. abs\/2211.05994 (2022)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-demo.26"},{"key":"e_1_3_2_1_14_1","volume-title":"Unifying Question Answering and Text Classification via Span Extraction. CoRR","author":"Keskar Nitish Shirish","year":"2019","unstructured":"Nitish Shirish Keskar , Bryan McCann , Caiming Xiong , and Richard Socher . 2019. Unifying Question Answering and Text Classification via Span Extraction. CoRR , Vol. abs\/ 1904 .09286 ( 2019 ). Nitish Shirish Keskar, Bryan McCann, Caiming Xiong, and Richard Socher. 2019. Unifying Question Answering and Text Classification via Span Extraction. CoRR, Vol. abs\/1904.09286 (2019)."},{"volume-title":"ACL","author":"Li Xiang Lisa","key":"e_1_3_2_1_15_1","unstructured":"Xiang Lisa Li and Percy Liang . 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation . In ACL , Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli (Eds.). Association for Computational Linguistics , 4582--4597. Xiang Lisa Li and Percy Liang. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In ACL, Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli (Eds.). Association for Computational Linguistics, 4582--4597."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3560815"},{"key":"e_1_3_2_1_17_1","volume-title":"abs\/2103.10385","author":"Liu Xiao","year":"2021","unstructured":"Xiao Liu , Yanan Zheng , Zhengxiao Du , Ming Ding , Yujie Qian , Zhilin Yang , and Jie Tang . 2021. GPT Understands , Too. Co RR , Vol . abs\/2103.10385 ( 2021 ). Xiao Liu, Yanan Zheng, Zhengxiao Du, Ming Ding, Yujie Qian, Zhilin Yang, and Jie Tang. 2021. GPT Understands, Too. CoRR, Vol. abs\/2103.10385 (2021)."},{"key":"e_1_3_2_1_18_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR , Vol. abs\/ 1907 .11692 ( 2019 ). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR, Vol. abs\/1907.11692 (2019)."},{"key":"e_1_3_2_1_19_1","volume-title":"Shengyu Fu, and Shujie Liu.","author":"Lu Shuai","year":"2021","unstructured":"Shuai Lu , Daya Guo , Shuo Ren , Junjie Huang , Alexey Svyatkovskiy , Ambrosio Blanco , Colin B. Clement , Dawn Drain , Daxin Jiang , Duyu Tang , Ge Li , Lidong Zhou , Linjun Shou , Long Zhou , Michele Tufano , Ming Gong , Ming Zhou , Nan Duan , Neel Sundaresan , Shao Kun Deng , Shengyu Fu, and Shujie Liu. 2021 . CodeXGLUE: A Machine Learning Benchmark Dataset for Code Understanding and Generation. CoRR , Vol. abs\/ 2102 .04664 (2021). Shuai Lu, Daya Guo, Shuo Ren, Junjie Huang, Alexey Svyatkovskiy, Ambrosio Blanco, Colin B. Clement, Dawn Drain, Daxin Jiang, Duyu Tang, Ge Li, Lidong Zhou, Linjun Shou, Long Zhou, Michele Tufano, Ming Gong, Ming Zhou, Nan Duan, Neel Sundaresan, Shao Kun Deng, Shengyu Fu, and Shujie Liu. 2021. CodeXGLUE: A Machine Learning Benchmark Dataset for Code Understanding and Generation. CoRR, Vol. abs\/2102.04664 (2021)."},{"volume-title":"UniPELT: A Unified Framework for Parameter-Efficient Language Model Tuning","author":"Mao Yuning","key":"e_1_3_2_1_20_1","unstructured":"Yuning Mao , Lambert Mathias , Rui Hou , Amjad Almahairi , Hao Ma , Jiawei Han , Scott Yih , and Madian Khabsa . 2022. UniPELT: A Unified Framework for Parameter-Efficient Language Model Tuning . In ACL. Association for Computational Linguistics , 6253--6264. Yuning Mao, Lambert Mathias, Rui Hou, Amjad Almahairi, Hao Ma, Jiawei Han, Scott Yih, and Madian Khabsa. 2022. UniPELT: A Unified Framework for Parameter-Efficient Language Model Tuning. In ACL. Association for Computational Linguistics, 6253--6264."},{"key":"e_1_3_2_1_21_1","volume-title":"Uncertainty-aware Self-training for Few-shot Text Classification. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020","author":"Mukherjee Subhabrata","year":"2020","unstructured":"Subhabrata Mukherjee and Ahmed Hassan Awadallah . 2020 . Uncertainty-aware Self-training for Few-shot Text Classification. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020 , NeurIPS 2020, December 6--12, 2020, virtual. Subhabrata Mukherjee and Ahmed Hassan Awadallah. 2020. Uncertainty-aware Self-training for Few-shot Text Classification. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6--12, 2020, virtual."},{"key":"e_1_3_2_1_22_1","volume-title":"Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models. CoRR","author":"Pan Xiaoman","year":"2022","unstructured":"Xiaoman Pan , Wenlin Yao , Hongming Zhang , Dian Yu , Dong Yu , and Jianshu Chen . 2022. Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models. CoRR , Vol. abs\/ 2210 .16433 ( 2022 ). Xiaoman Pan, Wenlin Yao, Hongming Zhang, Dian Yu, Dong Yu, and Jianshu Chen. 2022. Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models. CoRR, Vol. abs\/2210.16433 (2022)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3031898"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Minghui Qiu Peng Li Chengyu Wang Haojie Pan Ang Wang Cen Chen Xianyan Jia Yaliang Li Jun Huang Deng Cai and Wei Lin. 2021. EasyTransfer: A Simple and Scalable Deep Transfer Learning Platform for NLP Applications. In CIKM. ACM 4075--4084. Minghui Qiu Peng Li Chengyu Wang Haojie Pan Ang Wang Cen Chen Xianyan Jia Yaliang Li Jun Huang Deng Cai and Wei Lin. 2021. EasyTransfer: A Simple and Scalable Deep Transfer Learning Platform for NLP Applications. In CIKM. ACM 4075--4084.","DOI":"10.1145\/3459637.3481911"},{"key":"e_1_3_2_1_25_1","unstructured":"Alec Radford Jeffrey Wu Rewon Child David Luan Dario Amodei Ilya Sutskever etal 2019. Language models are unsupervised multitask learners. OpenAI blog Vol. 1 8 (2019) 9. Alec Radford Jeffrey Wu Rewon Child David Luan Dario Amodei Ilya Sutskever et al. 2019. Language models are unsupervised multitask learners. OpenAI blog Vol. 1 8 (2019) 9."},{"key":"e_1_3_2_1_26_1","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel , Noam Shazeer , Adam Roberts , Katherine Lee , Sharan Narang , Michael Matena , Yanqi Zhou , Wei Li , and Peter J. Liu . 2020 . Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer . J. Mach. Learn. Res. , Vol. 21 (2020), 140:1--140:67. Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res., Vol. 21 (2020), 140:1--140:67.","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Timo Schick and Hinrich Sch\u00fc tze. 2021. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. In EACL. 255--269. Timo Schick and Hinrich Sch\u00fc tze. 2021. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. In EACL. 255--269.","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"e_1_3_2_1_28_1","volume-title":"Zemel","author":"Snell Jake","year":"2017","unstructured":"Jake Snell , Kevin Swersky , and Richard S . Zemel . 2017 . Prototypical Networks for Few-shot Learning. In NIPS. 4077--4087. Jake Snell, Kevin Swersky, and Richard S. Zemel. 2017. Prototypical Networks for Few-shot Learning. In NIPS. 4077--4087."},{"key":"e_1_3_2_1_29_1","volume-title":"Global Pointer: Novel Efficient Span-based Approach for Named Entity Recognition. CoRR","author":"Su Jianlin","year":"2022","unstructured":"Jianlin Su , Ahmed Murtadha , Shengfeng Pan , Jing Hou , Jun Sun , Wanwei Huang , Bo Wen , and Yunfeng Liu . 2022 . Global Pointer: Novel Efficient Span-based Approach for Named Entity Recognition. CoRR , Vol. abs\/ 2208 .03054 (2022). Jianlin Su, Ahmed Murtadha, Shengfeng Pan, Jing Hou, Jun Sun, Wanwei Huang, Bo Wen, and Yunfeng Liu. 2022. Global Pointer: Novel Efficient Span-based Approach for Named Entity Recognition. CoRR, Vol. abs\/2208.03054 (2022)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"crossref","unstructured":"Austin Cheng-Yun Tsai Sheng-Ya Lin and Li-Chen Fu. 2022. Contrast-Enhanced Semi-supervised Text Classification with Few Labels. In AAAI. 11394--11402. Austin Cheng-Yun Tsai Sheng-Ya Lin and Li-Chen Fu. 2022. Contrast-Enhanced Semi-supervised Text Classification with Few Labels. In AAAI. 11394--11402.","DOI":"10.1609\/aaai.v36i10.21391"},{"volume-title":"GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In EMNLP Workshop BlackboxNLP.","author":"Wang Alex","key":"e_1_3_2_1_31_1","unstructured":"Alex Wang , Amanpreet Singh , Julian Michael , Felix Hill , Omer Levy , and Samuel R. Bowman . 2018 . GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In EMNLP Workshop BlackboxNLP. Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. 2018. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In EMNLP Workshop BlackboxNLP."},{"key":"e_1_3_2_1_32_1","volume-title":"EasyNLP: A Comprehensive and Easy-to-use Toolkit for Natural Language Processing. CoRR","author":"Wang Chengyu","year":"2022","unstructured":"Chengyu Wang , Minghui Qiu , Taolin Zhang , Tingting Liu , Lei Li , Jianing Wang , Ming Wang , Jun Huang , and Wei Lin . 2022c. EasyNLP: A Comprehensive and Easy-to-use Toolkit for Natural Language Processing. CoRR , Vol. abs\/ 2205 .00258 ( 2022 ). Chengyu Wang, Minghui Qiu, Taolin Zhang, Tingting Liu, Lei Li, Jianing Wang, Ming Wang, Jun Huang, and Wei Lin. 2022c. EasyNLP: A Comprehensive and Easy-to-use Toolkit for Natural Language Processing. CoRR, Vol. abs\/2205.00258 (2022)."},{"volume-title":"Knowledge Prompting in Pre-trained Language Model for Natural Language Understanding","author":"Wang Jianing","key":"e_1_3_2_1_33_1","unstructured":"Jianing Wang , Wenkang Huang , Minghui Qiu , Qiuhui Shi , Hongbin Wang , Xiang Li , and Ming Gao . 2022a. Knowledge Prompting in Pre-trained Language Model for Natural Language Understanding . In EMNLP. Association for Computational Linguistics , 3164--3177. Jianing Wang, Wenkang Huang, Minghui Qiu, Qiuhui Shi, Hongbin Wang, Xiang Li, and Ming Gao. 2022a. Knowledge Prompting in Pre-trained Language Model for Natural Language Understanding. In EMNLP. Association for Computational Linguistics, 3164--3177."},{"key":"e_1_3_2_1_34_1","volume-title":"2023 b. Uncertainty-aware Self-training for Low-resource Neural Sequence Labeling. CoRR","author":"Wang Jianing","year":"2023","unstructured":"Jianing Wang , Chengyu Wang , Jun Huang , Ming Gao , and Aoying Zhou . 2023 b. Uncertainty-aware Self-training for Low-resource Neural Sequence Labeling. CoRR , Vol. abs\/ 2302 .08659 ( 2023 ). Jianing Wang, Chengyu Wang, Jun Huang, Ming Gao, and Aoying Zhou. 2023 b. Uncertainty-aware Self-training for Low-resource Neural Sequence Labeling. CoRR, Vol. abs\/2302.08659 (2023)."},{"key":"e_1_3_2_1_35_1","volume-title":"Entailment as Few-Shot Learner. CoRR","author":"Wang Sinong","year":"2021","unstructured":"Sinong Wang , Han Fang , Madian Khabsa , Hanzi Mao , and Hao Ma. 2021a. Entailment as Few-Shot Learner. CoRR , Vol. abs\/ 2104 .14690 ( 2021 ). Sinong Wang, Han Fang, Madian Khabsa, Hanzi Mao, and Hao Ma. 2021a. Entailment as Few-Shot Learner. CoRR, Vol. abs\/2104.14690 (2021)."},{"key":"e_1_3_2_1_36_1","volume-title":"Nghi D.Q. Bui, Junnan Li, and Steven C. H. Hoi. 2023 a. CodeT5: Open Code Large Language Models for Code Understanding and Generation. arXiv preprint","author":"Wang Yue","year":"2023","unstructured":"Yue Wang , Hung Le , Akhilesh Deepak Gotmare , Nghi D.Q. Bui, Junnan Li, and Steven C. H. Hoi. 2023 a. CodeT5: Open Code Large Language Models for Code Understanding and Generation. arXiv preprint ( 2023 ). Yue Wang, Hung Le, Akhilesh Deepak Gotmare, Nghi D.Q. Bui, Junnan Li, and Steven C. H. Hoi. 2023 a. CodeT5: Open Code Large Language Models for Code Understanding and Generation. arXiv preprint (2023)."},{"key":"e_1_3_2_1_37_1","volume-title":"Ahmed Hassan Awadallah, and Jianfeng Gao","author":"Wang Yaqing","year":"2022","unstructured":"Yaqing Wang , Subhabrata Mukherjee , Xiaodong Liu , Jing Gao , Ahmed Hassan Awadallah, and Jianfeng Gao . 2022 b. LiST: Lite Prompted Self-training Makes Parameter-efficient Few-shot Learners. In NAACL. 2262--2281. Yaqing Wang, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, and Jianfeng Gao. 2022b. LiST: Lite Prompted Self-training Makes Parameter-efficient Few-shot Learners. In NAACL. 2262--2281."},{"key":"e_1_3_2_1_38_1","volume-title":"Hoi","author":"Wang Yue","year":"2021","unstructured":"Yue Wang , Weishi Wang , Shafiq R. Joty , and Steven C. H . Hoi . 2021 b. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. In EMNLP. Association for Computational Linguistics , 8696--8708. Yue Wang, Weishi Wang, Shafiq R. Joty, and Steven C. H. Hoi. 2021b. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. In EMNLP. Association for Computational Linguistics, 8696--8708."},{"key":"e_1_3_2_1_39_1","volume-title":"Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei , Maarten Bosma , Vincent Y. Zhao , Kelvin Guu , Adams Wei Yu , Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le. 2022 a. Finetuned Language Models are Zero-Shot Learners. In ICLR. OpenReview .net. Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le. 2022a. Finetuned Language Models are Zero-Shot Learners. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_40_1","volume-title":"Fei Xia, Ed H. Chi, Quoc V Le, and Denny Zhou.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei , Xuezhi Wang , Dale Schuurmans , Maarten Bosma , brian ichter , Fei Xia, Ed H. Chi, Quoc V Le, and Denny Zhou. 2022 b. Chain of Thought Prompting Elicits Reasoning in Large Language Models. In Advances in Neural Information Processing Systems, Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (Eds .). https:\/\/openreview.net\/forum?id=_VjQlMeSB_J Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, brian ichter, Fei Xia, Ed H. Chi, Quoc V Le, and Denny Zhou. 2022b. Chain of Thought Prompting Elicits Reasoning in Large Language Models. In Advances in Neural Information Processing Systems, Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (Eds.). https:\/\/openreview.net\/forum?id=_VjQlMeSB_J"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-demo.47"},{"key":"e_1_3_2_1_42_1","volume-title":"CLUE: A Chinese Language Understanding Evaluation Benchmark. In COLING, Donia Scott, N\u00fa ria Bel, and Chengqing Zong (Eds.). 4762--4772.","author":"Xu Liang","year":"2020","unstructured":"Liang Xu , Hai Hu , Xuanwei Zhang , Lu Li , Chenjie Cao , Yudong Li , Yechen Xu , Kai Sun , Dian Yu , Cong Yu , Yin Tian , Qianqian Dong , Weitang Liu , Bo Shi , Yiming Cui , Junyi Li , Jun Zeng , Rongzhao Wang , Weijian Xie , Yanting Li , Yina Patterson , Zuoyu Tian , Yiwen Zhang , He Zhou , Shaoweihua Liu , Zhe Zhao , Qipeng Zhao , Cong Yue , Xinrui Zhang , Zhengliang Yang , Kyle Richardson , and Zhenzhong Lan . 2020 . CLUE: A Chinese Language Understanding Evaluation Benchmark. In COLING, Donia Scott, N\u00fa ria Bel, and Chengqing Zong (Eds.). 4762--4772. Liang Xu, Hai Hu, Xuanwei Zhang, Lu Li, Chenjie Cao, Yudong Li, Yechen Xu, Kai Sun, Dian Yu, Cong Yu, Yin Tian, Qianqian Dong, Weitang Liu, Bo Shi, Yiming Cui, Junyi Li, Jun Zeng, Rongzhao Wang, Weijian Xie, Yanting Li, Yina Patterson, Zuoyu Tian, Yiwen Zhang, He Zhou, Shaoweihua Liu, Zhe Zhao, Qipeng Zhao, Cong Yue, Xinrui Zhang, Zhengliang Yang, Kyle Richardson, and Zhenzhong Lan. 2020. CLUE: A Chinese Language Understanding Evaluation Benchmark. In COLING, Donia Scott, N\u00fa ria Bel, and Chengqing Zong (Eds.). 4762--4772."},{"key":"e_1_3_2_1_43_1","volume-title":"Le","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang , Zihang Dai , Yiming Yang , Jaime G. Carbonell , Ruslan Salakhutdinov , and Quoc V . Le . 2019 . XLNet: Generalized Autoregressive Pretraining for Language Understanding. In NeurIPS. 5754--5764. Zhilin Yang, Zihang Dai, Yiming Yang, Jaime G. Carbonell, Ruslan Salakhutdinov, and Quoc V. Le. 2019. XLNet: Generalized Autoregressive Pretraining for Language Understanding. In NeurIPS. 5754--5764."},{"volume-title":"BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models","author":"Zaken Elad Ben","key":"e_1_3_2_1_44_1","unstructured":"Elad Ben Zaken , Yoav Goldberg , and Shauli Ravfogel . 2022. BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models . In ACL. Association for Computational Linguistics , 1--9. Elad Ben Zaken, Yoav Goldberg, and Shauli Ravfogel. 2022. BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models. In ACL. Association for Computational Linguistics, 1--9."},{"key":"e_1_3_2_1_45_1","volume-title":"DKPLM: Decomposable Knowledge-Enhanced Pre-trained Language Model for Natural Language Understanding","author":"Zhang Taolin","year":"2022","unstructured":"Taolin Zhang , Chengyu Wang , Nan Hu , Minghui Qiu , Chengguang Tang , Xiaofeng He , and Jun Huang . 2022 . DKPLM: Decomposable Knowledge-Enhanced Pre-trained Language Model for Natural Language Understanding . In AAAI. AAAI Press , 11703--11711. Taolin Zhang, Chengyu Wang, Nan Hu, Minghui Qiu, Chengguang Tang, Xiaofeng He, and Jun Huang. 2022. DKPLM: Decomposable Knowledge-Enhanced Pre-trained Language Model for Natural Language Understanding. In AAAI. AAAI Press, 11703--11711."},{"key":"e_1_3_2_1_46_1","volume-title":"ICML (Proceedings of Machine Learning Research","volume":"12706","author":"Zhao Zihao","year":"2021","unstructured":"Zihao Zhao , Eric Wallace , Shi Feng , Dan Klein , and Sameer Singh . 2021 . Calibrate Before Use: Improving Few-shot Performance of Language Models . In ICML (Proceedings of Machine Learning Research , Vol. 139). PMLR, 12697-- 12706 . Zihao Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate Before Use: Improving Few-shot Performance of Language Models. In ICML (Proceedings of Machine Learning Research, Vol. 139). PMLR, 12697--12706."}],"event":{"name":"CIKM '23: The 32nd ACM International Conference on Information and Knowledge Management","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web","SIGIR ACM Special Interest Group on Information Retrieval"],"location":"Birmingham United Kingdom","acronym":"CIKM '23"},"container-title":["Proceedings of the 32nd ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3583780.3614742","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3583780.3614742","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:46:30Z","timestamp":1750178790000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3583780.3614742"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,21]]},"references-count":46,"alternative-id":["10.1145\/3583780.3614742","10.1145\/3583780"],"URL":"https:\/\/doi.org\/10.1145\/3583780.3614742","relation":{},"subject":[],"published":{"date-parts":[[2023,10,21]]},"assertion":[{"value":"2023-10-21","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}