{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,29]],"date-time":"2025-09-29T12:40:11Z","timestamp":1759149611395,"version":"3.44.0"},"reference-count":25,"publisher":"Association for Computing Machinery (ACM)","issue":"10","funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62476007,U23B2029"],"award-info":[{"award-number":["62476007,U23B2029"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Asian Low-Resour. Lang. Inf. Process."],"published-print":{"date-parts":[[2025,10,31]]},"abstract":"<jats:p>The goal of continuous few-shot relation extraction is to enable the model to continuously learn new relation types under conditions with limited labeled training data while avoiding the forgetting of previously learned relations. The primary challenges include catastrophic forgetting of old relations and overfitting due to data sparsity. To address these challenges, this article proposes a hierarchical distillation model that innovatively combines contrastive distillation with orthogonal adversarial distillation techniques. Specifically, we introduce a contrastive distillation approach in the feature distillation layer, integrating adaptive cosine techniques and negative sampling strategies to ensure that the model effectively retains and utilizes knowledge from previous tasks when learning new ones. Additionally, we employ orthogonal adversarial distillation in the hidden distillation layer to alleviate overfitting in low-resource scenarios. Experimental results demonstrate that our proposed method significantly outperforms the current state-of-the-art models for continuous few-shot relation extraction on two benchmark datasets, validating its effectiveness in handling few-shot data and knowledge transfer.<\/jats:p>","DOI":"10.1145\/3765749","type":"journal-article","created":{"date-parts":[[2025,9,3]],"date-time":"2025-09-03T11:05:41Z","timestamp":1756897541000},"page":"1-13","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Continuous Adaptive Knowledge Distillation for Few-Shot Relation Extraction"],"prefix":"10.1145","volume":"24","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-3640-7285","authenticated-orcid":false,"given":"Shuo","family":"Zhao","sequence":"first","affiliation":[{"name":"School of Information, North China University of Technology","place":["Beijing, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2244-3764","authenticated-orcid":false,"given":"Jianyong","family":"Duan","sequence":"additional","affiliation":[{"name":"North China University of Technology","place":["Beijing, China"]}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3068-735X","authenticated-orcid":false,"given":"Li","family":"He","sequence":"additional","affiliation":[{"name":"North China University of Technology","place":["Beijing, China"]},{"name":"CNONIX National Standard Application and Promotion Lab","place":["Beijing, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0896-080X","authenticated-orcid":false,"given":"Hao","family":"Wang","sequence":"additional","affiliation":[{"name":"North China University of Technology","place":["Beijing, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5368-339X","authenticated-orcid":false,"given":"Qing","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Information Science and Technology, CNONIX National Standard Application and Promotion Laboratory, North China University of Technology","place":["Beijing, China"]}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5953-4566","authenticated-orcid":false,"given":"Jie","family":"Liu","sequence":"additional","affiliation":[{"name":"Information Engineering College, Capital Normal University","place":["Beijing, China"]}]}],"member":"320","published-online":{"date-parts":[[2025,9,29]]},"reference":[{"doi-asserted-by":"publisher","key":"e_1_3_1_2_2","DOI":"10.18653\/v1\/2024.naacl-short.6"},{"doi-asserted-by":"publisher","key":"e_1_3_1_3_2","DOI":"10.18653\/v1\/2023.acl-long.409"},{"doi-asserted-by":"publisher","key":"e_1_3_1_4_2","DOI":"10.18653\/v1\/2021.acl-long.20"},{"doi-asserted-by":"publisher","key":"e_1_3_1_5_2","DOI":"10.18653\/v1\/N19-1423"},{"doi-asserted-by":"publisher","key":"e_1_3_1_6_2","DOI":"10.1016\/S1364-6613(99)01294-2"},{"doi-asserted-by":"publisher","key":"e_1_3_1_7_2","DOI":"10.18653\/v1\/2021.emnlp-main.552"},{"doi-asserted-by":"publisher","key":"e_1_3_1_8_2","DOI":"10.18653\/v1\/D18-1514"},{"key":"e_1_3_1_9_2","first-page":"1885","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics","author":"Hu Chengwei","year":"2022","unstructured":"Chengwei Hu, Deqing Yang, Haoliang Jin, Zhen Chen, and Yanghua Xiao. 2022. Improving continual relation extraction through prototypical contrastive learning. In Proceedings of the 29th International Conference on Computational Linguistics, Nicoletta Calzolari, Chu-Ren Huang, Hansaem Kim, James Pustejovsky, Leo Wanner, Key-Sun Choi, Pum-Mo Ryu, Hsin-Hsi Chen, Lucia Donatelli, Heng Ji, Sadao Kurohashi, Patrizia Paggio, Nianwen Xue, Seokhwan Kim, Younggyun Hahm, Zhong He, Tony Kyungil Lee, Enrico Santus, Francis Bond, and Seung-Hoon Na (Eds.). International Committee on Computational Linguistics, Gyeongju, Republic of Korea, 1885\u20131895. Retrieved from https:\/\/aclanthology.org\/2022.coling-1.163\/"},{"doi-asserted-by":"publisher","key":"e_1_3_1_10_2","DOI":"10.18653\/v1\/2024.findings-acl.702"},{"doi-asserted-by":"publisher","key":"e_1_3_1_11_2","DOI":"10.18653\/v1\/2023.acl-long.852"},{"doi-asserted-by":"publisher","key":"e_1_3_1_12_2","DOI":"10.18653\/v1\/2023.findings-emnlp.710"},{"doi-asserted-by":"publisher","key":"e_1_3_1_13_2","DOI":"10.18653\/v1\/2020.emnlp-main.298"},{"doi-asserted-by":"publisher","key":"e_1_3_1_14_2","DOI":"10.18653\/v1\/2024.findings-acl.1"},{"doi-asserted-by":"publisher","key":"e_1_3_1_15_2","DOI":"10.18653\/v1\/2022.acl-long.198"},{"doi-asserted-by":"publisher","key":"e_1_3_1_16_2","DOI":"10.18653\/v1\/2024.acl-long.113"},{"doi-asserted-by":"publisher","key":"e_1_3_1_17_2","DOI":"10.18653\/v1\/2023.findings-emnlp.969"},{"doi-asserted-by":"publisher","key":"e_1_3_1_18_2","DOI":"10.18653\/v1\/2023.findings-acl.804"},{"doi-asserted-by":"publisher","key":"e_1_3_1_19_2","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"doi-asserted-by":"publisher","key":"e_1_3_1_20_2","DOI":"10.18653\/v1\/2023.findings-acl.638"},{"key":"e_1_3_1_21_2","first-page":"2494","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics","author":"Yao Yuanzhou","year":"2022","unstructured":"Yuanzhou Yao, Zhao Zhang, Yongjun Xu, and Chao Li. 2022. Data augmentation for few-shot knowledge graph completion from hierarchical perspective. In Proceedings of the 29th International Conference on Computational Linguistics, Nicoletta Calzolari, Chu-Ren Huang, Hansaem Kim, James Pustejovsky, Leo Wanner, Key-Sun Choi, Pum-Mo Ryu, Hsin-Hsi Chen, Lucia Donatelli, Heng Ji, Sadao Kurohashi, Patrizia Paggio, Nianwen Xue, Seokhwan Kim, Younggyun Hahm, Zhong He, Tony Kyungil Lee, Enrico Santus, Francis Bond, and Seung-Hoon Na (Eds.). International Committee on Computational Linguistics, Gyeongju, Republic of Korea, 2494\u20132503. Retrieved from https:\/\/aclanthology.org\/2022.coling-1.220\/"},{"doi-asserted-by":"publisher","key":"e_1_3_1_22_2","DOI":"10.18653\/v1\/2024.acl-long.603"},{"doi-asserted-by":"publisher","key":"e_1_3_1_23_2","DOI":"10.18653\/v1\/D17-1004"},{"key":"e_1_3_1_24_2","first-page":"8756","volume-title":"Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)","author":"Zhang Zhiheng","year":"2024","unstructured":"Zhiheng Zhang, Daojian Zeng, and Xue Bai. 2024. Improving continual few-shot relation extraction through relational knowledge distillation and prototype augmentation. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), Nicoletta Calzolari, Min-Yen Kan, Veronique Hoste, Alessandro Lenci, Sakriani Sakti, and Nianwen Xue (Eds.). ELRA and ICCL, Torino, Italia, 8756\u20138767. Retrieved from https:\/\/aclanthology.org\/2024.lrec-main.767\/"},{"doi-asserted-by":"publisher","key":"e_1_3_1_25_2","DOI":"10.18653\/v1\/2022.findings-acl.268"},{"doi-asserted-by":"publisher","key":"e_1_3_1_26_2","DOI":"10.18653\/v1\/2023.acl-long.65"}],"container-title":["ACM Transactions on Asian and Low-Resource Language Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3765749","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,29]],"date-time":"2025-09-29T12:21:45Z","timestamp":1759148505000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3765749"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,29]]},"references-count":25,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,10,31]]}},"alternative-id":["10.1145\/3765749"],"URL":"https:\/\/doi.org\/10.1145\/3765749","relation":{},"ISSN":["2375-4699","2375-4702"],"issn-type":[{"type":"print","value":"2375-4699"},{"type":"electronic","value":"2375-4702"}],"subject":[],"published":{"date-parts":[[2025,9,29]]},"assertion":[{"value":"2024-11-14","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-08-26","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-09-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}