{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T16:11:30Z","timestamp":1774541490001,"version":"3.50.1"},"reference-count":62,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2024,4,26]],"date-time":"2024-04-26T00:00:00Z","timestamp":1714089600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2022YFF0902100"],"award-info":[{"award-number":["2022YFF0902100"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62376262"],"award-info":[{"award-number":["62376262"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100017610","name":"Shenzhen Science and Technology Innovation Program","doi-asserted-by":"crossref","award":["KQTD20190929172835662"],"award-info":[{"award-number":["KQTD20190929172835662"]}],"id":[{"id":"10.13039\/501100017610","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Shenzhen Basic Research Foundation","award":["JCYJ20210324115 614039 and JCYJ20200109113441941"],"award-info":[{"award-number":["JCYJ20210324115 614039 and JCYJ20200109113441941"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Inf. Syst."],"published-print":{"date-parts":[[2024,7,31]]},"abstract":"<jats:p>\n            The goal of semi-supervised text classification (SSTC) is to train a model by exploring both a small number of labeled data and a large number of unlabeled data, such that the learned semi-supervised classifier performs better than the supervised classifier trained on solely the labeled samples. Pseudo-labeling is one of the most widely used SSTC techniques, which trains a teacher classifier with a small number of labeled examples to predict pseudo labels for the unlabeled data. The generated pseudo-labeled examples are then utilized to train a student classifier, such that the learned student classifier can outperform the teacher classifier. Nevertheless, the predicted pseudo labels may be inaccurate, making the performance of the student classifier degraded. The student classifier may perform even worse than the teacher classifier. To alleviate this issue, in this paper, we introduce a dual meta-learning (\n            <jats:bold>DML<\/jats:bold>\n            ) technique for semi-supervised text classification, which improves the teacher and student classifiers simultaneously in an iterative manner. Specifically, we propose a meta-noise correction method to improve the student classifier by proposing a Noise Transition Matrix (NTM) with meta-learning to rectify the noisy pseudo labels. In addition, we devise a meta pseudo supervision method to improve the teacher classifier. Concretely, we exploit the feedback performance from the student classifier to further guide the teacher classifier to produce more accurate pseudo labels for the unlabeled data. In this way, both teacher and student classifiers can co-evolve in the iterative training process. Extensive experiments on four benchmark datasets highlight the effectiveness of our DML method against existing state-of-the-art methods for semi-supervised text classification. We release our code and data of this paper publicly at https:\/\/github.com\/GRIT621\/DML.\n          <\/jats:p>","DOI":"10.1145\/3648612","type":"journal-article","created":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T12:30:03Z","timestamp":1708432203000},"page":"1-28","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["Improving Semi-Supervised Text Classification with Dual Meta-Learning"],"prefix":"10.1145","volume":"42","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5239-2454","authenticated-orcid":false,"given":"Shujie","family":"Li","sequence":"first","affiliation":[{"name":"Shenzhen Institutes of Advanced Technology, Chinese Academy of Science,  Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1559-2996","authenticated-orcid":false,"given":"Guanghu","family":"Yuan","sequence":"additional","affiliation":[{"name":"Shenzhen Institutes of Advanced Technology, Chinese Academy of Science,  Shenzhen China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7345-5071","authenticated-orcid":false,"given":"Min","family":"Yang","sequence":"additional","affiliation":[{"name":"Shenzhen Institutes of Advanced Technology Chinese Academy of Sciences,  Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3220-904X","authenticated-orcid":false,"given":"Ying","family":"Shen","sequence":"additional","affiliation":[{"name":"School of Intelligent Systems Engineering, Sun Yat-Sen University,  Guangdong China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4592-3875","authenticated-orcid":false,"given":"Chengming","family":"Li","sequence":"additional","affiliation":[{"name":"School of Intelligent Systems Engineering, Sun Yat-Sen University,  Guangdong China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9885-2364","authenticated-orcid":false,"given":"Ruifeng","family":"Xu","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology (Shenzhen), Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6001-1260","authenticated-orcid":false,"given":"Xiaoyan","family":"Zhao","sequence":"additional","affiliation":[{"name":"Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences,  Shenzhen, China"}]}],"member":"320","published-online":{"date-parts":[[2024,4,26]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403214"},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403264"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1145\/3469856"},{"key":"e_1_3_2_5_2","article-title":"Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results","author":"Tarvainen Antti","year":"2017","unstructured":"Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. NIPS (2017).","journal-title":"NIPS"},{"key":"e_1_3_2_6_2","first-page":"1314","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence","volume":"36","author":"Li Hengduo","year":"2022","unstructured":"Hengduo Li, Zuxuan Wu, Abhinav Shrivastava, and Larry S. Davis. 2022. Rethinking pseudo labels for semi-supervised object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36. 1314\u20131322."},{"key":"e_1_3_2_7_2","first-page":"18408","article-title":"FlexMatch: Boosting semi-supervised learning with curriculum pseudo labeling","volume":"34","author":"Zhang Bowen","year":"2021","unstructured":"Bowen Zhang, Yidong Wang, Wenxin Hou, Hao Wu, Jindong Wang, Manabu Okumura, and Takahiro Shinozaki. 2021. FlexMatch: Boosting semi-supervised learning with curriculum pseudo labeling. Advances in Neural Information Processing Systems 34 (2021), 18408\u201318419.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_8_2","article-title":"MixText: Linguistically-informed interpolation of hidden space for semi-supervised text classification","author":"Chen Jiaao","year":"2020","unstructured":"Jiaao Chen, Zichao Yang, and Diyi Yang. 2020. MixText: Linguistically-informed interpolation of hidden space for semi-supervised text classification. ACL (2020).","journal-title":"ACL"},{"key":"e_1_3_2_9_2","article-title":"Are anchor points really indispensable in label-noise learning?","volume":"32","author":"Xia Xiaobo","year":"2019","unstructured":"Xiaobo Xia, Tongliang Liu, Nannan Wang, Bo Han, Chen Gong, Gang Niu, and Masashi Sugiyama. 2019. Are anchor points really indispensable in label-noise learning? Advances in Neural Information Processing Systems 32 (2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_10_2","article-title":"CLC: A consensus-based label correction approach in federated learning","author":"Zeng Bixiao","year":"2022","unstructured":"Bixiao Zeng, Xiaodong Yang, Yiqiang Chen, Hanchao Yu, and Yingwei Zhang. 2022. CLC: A consensus-based label correction approach in federated learning. ACM Transactions on Intelligent Systems and Technology (TIST) (2022).","journal-title":"ACM Transactions on Intelligent Systems and Technology (TIST)"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1145\/3450352"},{"key":"e_1_3_2_12_2","first-page":"60","volume-title":"Proceedings of the International AAAI Conference on Web and Social Media","volume":"14","author":"Bozarth Lia","year":"2020","unstructured":"Lia Bozarth and Ceren Budak. 2020. Toward a better performance evaluation framework for fake news classification. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 14. 60\u201371."},{"key":"e_1_3_2_13_2","first-page":"6280","volume-title":"Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing","author":"Jiang Qingnan","year":"2019","unstructured":"Qingnan Jiang, Lei Chen, Ruifeng Xu, Xiang Ao, and Min Yang. 2019. A challenge dataset and effective models for aspect-based sentiment analysis. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. 6280\u20136285."},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531784"},{"key":"e_1_3_2_15_2","first-page":"784","article-title":"An attention-based unsupervised adversarial model for movie review spam detection","volume":"23","author":"Gao Yuan","year":"2020","unstructured":"Yuan Gao, Maoguo Gong, Yu Xie, and Alex Kai Qin. 2020. An attention-based unsupervised adversarial model for movie review spam detection. IEEE Transactions on Multimedia 23 (2020), 784\u2013796.","journal-title":"IEEE Transactions on Multimedia"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531875"},{"key":"e_1_3_2_17_2","article-title":"Variational pretraining for semi-supervised text classification","author":"Gururangan Suchin","year":"2019","unstructured":"Suchin Gururangan, Tam Dang, Dallas Card, and Noah A. Smith. 2019. Variational pretraining for semi-supervised text classification. arXiv preprint arXiv:1906.02242 (2019).","journal-title":"arXiv preprint arXiv:1906.02242"},{"key":"e_1_3_2_18_2","article-title":"Semi-supervised sequence modeling with cross-view training","author":"Clark Kevin","year":"2018","unstructured":"Kevin Clark, Minh-Thang Luong, Christopher D. Manning, and Quoc V. Le. 2018. Semi-supervised sequence modeling with cross-view training. arXiv preprint arXiv:1809.08370 (2018).","journal-title":"arXiv preprint arXiv:1809.08370"},{"key":"e_1_3_2_19_2","first-page":"896","volume-title":"Workshop on Challenges in Representation Learning, ICML","author":"Lee Dong-Hyun","year":"2013","unstructured":"Dong-Hyun Lee. 2013. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on Challenges in Representation Learning, ICML, Vol. 3. 896."},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207304"},{"key":"e_1_3_2_21_2","article-title":"Curriculum labeling: Revisiting pseudo-labeling for semi-supervised learning","author":"Cascante-Bonilla Paola","year":"2020","unstructured":"Paola Cascante-Bonilla, Fuwen Tan, Yanjun Qi, and Vicente Ordonez. 2020. Curriculum labeling: Revisiting pseudo-labeling for semi-supervised learning. arXiv preprint arXiv:2001.06001 (2020).","journal-title":"arXiv preprint arXiv:2001.06001"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01219-9_18"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_19"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01395-y"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1145\/2516889"},{"key":"e_1_3_2_26_2","first-page":"12501","volume-title":"International Conference on Machine Learning","author":"Zhang Yivan","year":"2021","unstructured":"Yivan Zhang, Gang Niu, and Masashi Sugiyama. 2021. Learning noise transition matrix from only noisy labels via total variation regularization. In International Conference on Machine Learning. PMLR, 12501\u201312512."},{"key":"e_1_3_2_27_2","article-title":"Identifiability of label noise transition matrix.","author":"Liu Yang","year":"2022","unstructured":"Yang Liu. 2022. Identifiability of label noise transition matrix. arXiv preprint arXiv:2202.02016 (2022).","journal-title":"arXiv preprint arXiv:2202.02016"},{"issue":"7","key":"e_1_3_2_28_2","doi-asserted-by":"crossref","first-page":"2387","DOI":"10.1109\/TNNLS.2019.2935608","article-title":"Domain adaptation with neural embedding matching","volume":"31","author":"Wang Zengmao","year":"2019","unstructured":"Zengmao Wang, Bo Du, and Yuhong Guo. 2019. Domain adaptation with neural embedding matching. IEEE Transactions on Neural Networks and Learning Systems 31, 7 (2019), 2387\u20132397.","journal-title":"IEEE Transactions on Neural Networks and Learning Systems"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.1145\/2168752.2168767"},{"key":"e_1_3_2_30_2","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2012.2214488"},{"key":"e_1_3_2_31_2","doi-asserted-by":"crossref","first-page":"991","DOI":"10.1145\/3357384.3358040","volume-title":"Proceedings of the 28th ACM International Conference on Information and Knowledge Management","author":"Rizos Georgios","year":"2019","unstructured":"Georgios Rizos, Konstantin Hemker, and Bj\u00f6rn Schuller. 2019. Augment to prevent: Short-text data augmentation in deep learning for hate-speech classification. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management. 991\u20131000."},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401308"},{"key":"e_1_3_2_33_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00685"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1145\/3568396"},{"key":"e_1_3_2_35_2","article-title":"Temporal ensembling for semi-supervised learning","author":"Laine Samuli","year":"2017","unstructured":"Samuli Laine and Timo Aila. 2017. Temporal ensembling for semi-supervised learning. ICLR (2017).","journal-title":"ICLR"},{"issue":"1","key":"e_1_3_2_36_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3414067","article-title":"Graph-based regularization on embedding layers for recommendation","volume":"39","author":"Zhang Yuan","year":"2020","unstructured":"Yuan Zhang, Fei Sun, Xiaoyong Yang, Chen Xu, Wenwu Ou, and Yan Zhang. 2020. Graph-based regularization on embedding layers for recommendation. ACM Transactions on Information Systems (TOIS) 39, 1 (2020), 1\u201327.","journal-title":"ACM Transactions on Information Systems (TOIS)"},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2858821"},{"key":"e_1_3_2_38_2","first-page":"6256","article-title":"Unsupervised data augmentation for consistency training","volume":"33","author":"Xie Qizhe","year":"2020","unstructured":"Qizhe Xie, Zihang Dai, Eduard Hovy, Thang Luong, and Quoc Le. 2020. Unsupervised data augmentation for consistency training. Advances in Neural Information Processing Systems 33 (2020), 6256\u20136268.","journal-title":"Advances in Neural Information Processing Systems"},{"issue":"4","key":"e_1_3_2_39_2","first-page":"1440","article-title":"Robust graph-based semisupervised learning for noisy labeled data via maximum correntropy criterion","volume":"49","author":"Du Bo","year":"2018","unstructured":"Bo Du, Tang Xinyao, Zengmao Wang, Lefei Zhang, and Dacheng Tao. 2018. Robust graph-based semisupervised learning for noisy labeled data via maximum correntropy criterion. IEEE Transactions on Cybernetics 49, 4 (2018), 1440\u20131453.","journal-title":"IEEE Transactions on Cybernetics"},{"key":"e_1_3_2_40_2","first-page":"1126","volume-title":"International Conference on Machine Learning","author":"Finn Chelsea","year":"2017","unstructured":"Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In International Conference on Machine Learning. PMLR, 1126\u20131135."},{"key":"e_1_3_2_41_2","first-page":"2304","volume-title":"International Conference on Machine Learning","author":"Jiang Lu","year":"2018","unstructured":"Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. 2018. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In International Conference on Machine Learning. PMLR, 2304\u20132313."},{"key":"e_1_3_2_42_2","article-title":"Meta-weight-net: Learning an explicit mapping for sample weighting","volume":"32","author":"Shu Jun","year":"2019","unstructured":"Jun Shu, Qi Xie, Lixuan Yi, Qian Zhao, Sanping Zhou, Zongben Xu, and Deyu Meng. 2019. Meta-weight-net: Learning an explicit mapping for sample weighting. Advances in Neural Information Processing Systems 32 (2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_43_2","article-title":"On first-order meta-learning algorithms","author":"Nichol Alex","year":"2018","unstructured":"Alex Nichol, Joshua Achiam, and John Schulman. 2018. On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018).","journal-title":"arXiv preprint arXiv:1803.02999"},{"key":"e_1_3_2_44_2","article-title":"Few-shot aspect category sentiment analysis via meta-learning","author":"Liang Bin","year":"2022","unstructured":"Bin Liang, Xiang Li, Lin Gui, Yonghao Fu, Yulan He, Min Yang, and Ruifeng Xu. 2022. Few-shot aspect category sentiment analysis via meta-learning. ACM Transactions on Information Systems (TOIS) (2022).","journal-title":"ACM Transactions on Information Systems (TOIS)"},{"key":"e_1_3_2_45_2","article-title":"Meta learning for natural language processing: A survey","author":"Lee Hung-yi","year":"2022","unstructured":"Hung-yi Lee, Shang-Wen Li, and Ngoc Thang Vu. 2022. Meta learning for natural language processing: A survey. arXiv preprint arXiv:2205.01500 (2022).","journal-title":"arXiv preprint arXiv:2205.01500"},{"key":"e_1_3_2_46_2","article-title":"Learning to learn and predict: A meta-learning approach for multi-label classification","author":"Wu Jiawei","year":"2019","unstructured":"Jiawei Wu, Wenhan Xiong, and William Yang Wang. 2019. Learning to learn and predict: A meta-learning approach for multi-label classification. arXiv preprint arXiv:1909.04176 (2019).","journal-title":"arXiv preprint arXiv:1909.04176"},{"key":"e_1_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01139"},{"key":"e_1_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1145\/3488560.3498371"},{"key":"e_1_3_2_49_2","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539313"},{"key":"e_1_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531887"},{"key":"e_1_3_2_51_2","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. NAACL (2019).","journal-title":"NAACL"},{"key":"e_1_3_2_52_2","first-page":"649","article-title":"Character-level convolutional networks for text classification","volume":"28","author":"Zhang Xiang","year":"2015","unstructured":"Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. Advances in Neural Information Processing Systems 28 (2015), 649\u2013657.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_53_2","first-page":"830","volume-title":"AAAI","author":"Chang Ming-Wei","year":"2008","unstructured":"Ming-Wei Chang, Lev-Arie Ratinov, Dan Roth, and Vivek Srikumar. 2008. Importance of semantic representation: Dataless classification. In AAAI, Vol. 2. 830\u2013835."},{"key":"e_1_3_2_54_2","article-title":"Unsupervised data augmentation for consistency training","author":"Xie Qizhe","year":"2019","unstructured":"Qizhe Xie, Zihang Dai, Eduard Hovy, Minh-Thang Luong, and Quoc V. Le. 2019. Unsupervised data augmentation for consistency training. The 34th Conference on Neural Information Processing Systems (2019).","journal-title":"The 34th Conference on Neural Information Processing Systems"},{"key":"e_1_3_2_55_2","first-page":"5044","volume-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics","author":"Li Changchun","year":"2021","unstructured":"Changchun Li, Ximing Li, and Jihong Ouyang. 2021. Semi-supervised text classification with balanced deep representation distributions. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics. 5044\u20135053."},{"key":"e_1_3_2_56_2","article-title":"Rank-aware negative training for semi-supervised text classification","author":"Murtadha Ahmed","year":"2023","unstructured":"Ahmed Murtadha, Shengfeng Pan, Wen Bo, Jianlin Su, Xinxin Cao, Wenze Zhang, and Yunfeng Liu. 2023. Rank-aware negative training for semi-supervised text classification. arXiv preprint arXiv:2306.07621 (2023).","journal-title":"arXiv preprint arXiv:2306.07621"},{"key":"e_1_3_2_57_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-35289-8_25"},{"key":"e_1_3_2_58_2","unstructured":"Aiwei Liu Xuming Hu Lijie Wen and Philip S. Yu. 2023. A comprehensive evaluation of ChatGPT\u2019s zero-shot Text-to-SQL capability. (2023). arxiv:cs.CL\/2303.13547"},{"key":"e_1_3_2_59_2","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Dan Bikel Lukas Blecher Cristian Canton Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel Kloumann Artem Korenev Punit Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric Michael Smith Ranjan Subramanian Xiaoqing Ellen Tan Binh Tang Ross Taylor Adina Williams Jian Xiang Kuan Puxin Xu Zheng Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov and Thomas Scialom. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_60_2","article-title":"Speak, memory: An archaeology of books known to ChatGPT\/GPT-4","author":"Chang Kent K.","year":"2023","unstructured":"Kent K. Chang, Mackenzie Cramer, Sandeep Soni, and David Bamman. 2023. Speak, memory: An archaeology of books known to ChatGPT\/GPT-4. arXiv preprint arXiv:2305.00118 (2023).","journal-title":"arXiv preprint arXiv:2305.00118"},{"key":"e_1_3_2_61_2","article-title":"LLMs as counterfactual explanation modules: Can ChatGPT explain black-box text classifiers?","author":"Bhattacharjee Amrita","year":"2023","unstructured":"Amrita Bhattacharjee, Raha Moraffah, Joshua Garland, and Huan Liu. 2023. LLMs as counterfactual explanation modules: Can ChatGPT explain black-box text classifiers? arXiv preprint arXiv:2309.13340 (2023).","journal-title":"arXiv preprint arXiv:2309.13340"},{"key":"e_1_3_2_62_2","article-title":"Learning with symmetric label noise: The importance of being unhinged","volume":"28","author":"Rooyen Brendan Van","year":"2015","unstructured":"Brendan Van Rooyen, Aditya Menon, and Robert C. Williamson. 2015. Learning with symmetric label noise: The importance of being unhinged. Advances in Neural Information Processing Systems 28 (2015).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_63_2","first-page":"596","article-title":"FixMatch: Simplifying semi-supervised learning with consistency and confidence","volume":"33","author":"Sohn Kihyuk","year":"2020","unstructured":"Kihyuk Sohn, David Berthelot, Nicholas Carlini, Zizhao Zhang, Han Zhang, Colin A. Raffel, Ekin Dogus Cubuk, Alexey Kurakin, and Chun-Liang Li. 2020. FixMatch: Simplifying semi-supervised learning with consistency and confidence. Advances in Neural Information Processing Systems 33 (2020), 596\u2013608.","journal-title":"Advances in Neural Information Processing Systems"}],"container-title":["ACM Transactions on Information Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3648612","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3648612","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T22:50:20Z","timestamp":1750287020000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3648612"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,26]]},"references-count":62,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,7,31]]}},"alternative-id":["10.1145\/3648612"],"URL":"https:\/\/doi.org\/10.1145\/3648612","relation":{},"ISSN":["1046-8188","1558-2868"],"issn-type":[{"value":"1046-8188","type":"print"},{"value":"1558-2868","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4,26]]},"assertion":[{"value":"2022-11-15","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-02-06","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-04-26","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}