{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T14:10:10Z","timestamp":1749132610685,"version":"3.41.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T00:00:00Z","timestamp":1749081600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T00:00:00Z","timestamp":1749081600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Supercomput"],"DOI":"10.1007\/s11227-025-07475-5","type":"journal-article","created":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T13:33:13Z","timestamp":1749130393000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["ECKT: enhancing cross-task knowledge transfer in continual few-shot relation extraction"],"prefix":"10.1007","volume":"81","author":[{"given":"Yifan","family":"Yao","sequence":"first","affiliation":[]},{"given":"Juan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Zhijun","family":"Fang","sequence":"additional","affiliation":[]},{"given":"Yongbin","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,5]]},"reference":[{"key":"7475_CR1","doi-asserted-by":"publisher","unstructured":"Benaim S, Wolf L (2018) One-shot unsupervised cross domain translation. Advances in neural information processing systems 31. https:\/\/doi.org\/10.48550\/arXiv.1806.06029","DOI":"10.48550\/arXiv.1806.06029"},{"key":"7475_CR2","doi-asserted-by":"publisher","unstructured":"Chaudhry A, Ranzato M, Rohrbach M et\u00a0al (2019) Efficient lifelong learning with a-gem. International Conference on Learning Representations https:\/\/doi.org\/10.48550\/arXiv.1812.00420","DOI":"10.48550\/arXiv.1812.00420"},{"key":"7475_CR3","doi-asserted-by":"publisher","unstructured":"Chen T, Goodfellow I, Shlens J (2015) Net2net: accelerating learning via knowledge transfer. International Conference on Learning Representations https:\/\/doi.org\/10.48550\/arXiv.1511.05641","DOI":"10.48550\/arXiv.1511.05641"},{"key":"7475_CR4","doi-asserted-by":"publisher","unstructured":"Chen X, Zhang N, Xie X et al (2022) Knowprompt: knowledge-aware prompt-tuning with synergistic optimization for relation extraction. In: Proceedings of the ACM Web conference 2022:2778\u20132788. https:\/\/doi.org\/10.1145\/3485447.3511998","DOI":"10.1145\/3485447.3511998"},{"key":"7475_CR5","doi-asserted-by":"publisher","unstructured":"Chen X, Wu H, Shi X (2023) Consistent prototype learning for few-shot continual relation extraction. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 7409\u20137422, https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.409","DOI":"10.18653\/v1\/2023.acl-long.409"},{"key":"7475_CR6","doi-asserted-by":"publisher","unstructured":"Cui L, Yang D, Yu J et\u00a0al (2021) Refining sample embeddings with relation prototypes to enhance continual relation extraction. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp 232\u2013243, https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.20","DOI":"10.18653\/v1\/2021.acl-long.20"},{"key":"7475_CR7","doi-asserted-by":"publisher","unstructured":"Devlin J (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics p 4171-4186. https:\/\/doi.org\/10.48550\/arXiv.1810.04805","DOI":"10.48550\/arXiv.1810.04805"},{"key":"7475_CR8","doi-asserted-by":"publisher","unstructured":"Finn C, Abbeel P, Levine S (2017a) Model-agnostic meta-learning for fast adaptation of deep networks. In: International Conference on Machine Learning, PMLR, pp 1126\u20131135, https:\/\/doi.org\/10.48550\/arXiv.1703.03400","DOI":"10.48550\/arXiv.1703.03400"},{"key":"7475_CR9","doi-asserted-by":"publisher","unstructured":"Finn C, Abbeel P, Levine S (2017b) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning, PMLR, pp 1126\u20131135, https:\/\/doi.org\/10.48550\/arXiv.1703.03400","DOI":"10.48550\/arXiv.1703.03400"},{"key":"7475_CR10","doi-asserted-by":"publisher","unstructured":"Fu Y, Xie Y, Fu Y, et\u00a0al (2022) Me-d2n: Multi-expert domain decompositional network for cross-domain few-shot learning. In: Proceedings of the 30th ACM International Conference on Multimedia, pp 6609\u20136617, https:\/\/doi.org\/10.48550\/arXiv.2210.05280","DOI":"10.48550\/arXiv.2210.05280"},{"key":"7475_CR11","doi-asserted-by":"publisher","unstructured":"Han X, Zhu H, Yu P, et\u00a0al (2018) Fewrel: a large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing pp 4803\u20134809. https:\/\/doi.org\/10.18653\/v1\/D18-1514","DOI":"10.18653\/v1\/D18-1514"},{"key":"7475_CR12","doi-asserted-by":"publisher","unstructured":"Han X, Dai Y, Gao T, et\u00a0al (2020) Continual relation learning via episodic memory activation and reconsolidation. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp 6429\u20136440, https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.573","DOI":"10.18653\/v1\/2020.acl-main.573"},{"key":"7475_CR13","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1016\/j.aiopen.2022.11.003","volume":"3","author":"X Han","year":"2022","unstructured":"Han X, Zhao W, Ding N et al (2022) Ptr: Prompt tuning with rules for text classification. AI Open 3:182\u2013192. https:\/\/doi.org\/10.1016\/j.aiopen.2022.11.003","journal-title":"AI Open"},{"key":"7475_CR14","doi-asserted-by":"publisher","unstructured":"Hoffman J, Tzeng E, Donahue J, et\u00a0al (2013) One-shot adaptation of supervised deep convolutional models. arXiv preprint arXiv:1312.6204https:\/\/doi.org\/10.48550\/arXiv.1312.6204","DOI":"10.48550\/arXiv.1312.6204"},{"key":"7475_CR15","doi-asserted-by":"publisher","unstructured":"Hu X, Zhang C, Ma F et\u00a0al (2021) Semi-supervised relation extraction via incremental meta self-training. In: Empirical Methods in Natural Language Processing, https:\/\/doi.org\/10.48550\/arXiv.2010.16410","DOI":"10.48550\/arXiv.2010.16410"},{"key":"7475_CR16","doi-asserted-by":"publisher","first-page":"166","DOI":"10.1016\/j.neucom.2022.07.079","volume":"507","author":"H Kang","year":"2022","unstructured":"Kang H, Li X, Jin L et al (2022) Tspnet: Translation supervised prototype network via residual learning for multimodal social relation extraction. Neurocomputing 507:166\u2013179. https:\/\/doi.org\/10.1016\/j.neucom.2022.07.079","journal-title":"Neurocomputing"},{"key":"7475_CR17","doi-asserted-by":"publisher","unstructured":"Kirkpatrick J, Pascanu R, Rabinowitz N et al (2017) Overcoming catastrophic forgetting in neural networks. In: Proceedings of the National Academy of Sciences 114(13):3521\u20133526. https:\/\/doi.org\/10.48550\/arXiv.1612.00796","DOI":"10.48550\/arXiv.1612.00796"},{"issue":"3","key":"7475_CR18","doi-asserted-by":"publisher","first-page":"2629","DOI":"10.1007\/s10489-022-03547-8","volume":"53","author":"F Li","year":"2023","unstructured":"Li F, Chen G, Liu X (2023) An event-based automatic annotation method for datasets of interpersonal relation extraction. Applied Intelligence 53(3):2629\u20132639. https:\/\/doi.org\/10.1007\/s10489-022-03547-8","journal-title":"Applied Intelligence"},{"key":"7475_CR19","doi-asserted-by":"publisher","unstructured":"Li W, Wang Q, Wu J et\u00a0al (2022) Piecewise convolutional neural networks with position attention and similar bag attention for distant supervision relation extraction. Applied Intelligence pp 1\u201311. https:\/\/doi.org\/10.1007\/s10489-021-02632-8","DOI":"10.1007\/s10489-021-02632-8"},{"issue":"12","key":"7475_CR20","doi-asserted-by":"publisher","first-page":"2935","DOI":"10.48550\/arXiv.1606.09282","volume":"40","author":"Z Li","year":"2017","unstructured":"Li Z, Hoiem D (2017) Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence 40(12):2935\u20132947. https:\/\/doi.org\/10.48550\/arXiv.1606.09282","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"7475_CR21","doi-asserted-by":"publisher","unstructured":"Liu J, Chen S, Wang B et\u00a0al (2021) Attention as relation: learning supervised multi-head self-attention for relation extraction. In: Proceedings of the twenty-ninth International Conference on International Joint Conferences on Artificial Intelligence, pp 3787\u20133793, https:\/\/doi.org\/10.24963\/ijcai.2020\/524","DOI":"10.24963\/ijcai.2020\/524"},{"key":"7475_CR22","doi-asserted-by":"publisher","unstructured":"Liu X, Masana M, Herranz L et\u00a0al (2018) Rotate your networks: Better weight consolidation and less catastrophic forgetting. In: 2018 24th International Conference on Pattern Recognition (ICPR), IEEE, pp 2262\u20132268, https:\/\/doi.org\/10.48550\/arXiv.1802.02950","DOI":"10.48550\/arXiv.1802.02950"},{"key":"7475_CR23","doi-asserted-by":"publisher","unstructured":"Ma, Han, Liang (2024) Making pre-trained language models better continual few-shot relation extractors. In: LREC-COLING 2024, pp 10970\u201310983, https:\/\/doi.org\/10.48550\/arXiv.2402.15713","DOI":"10.48550\/arXiv.2402.15713"},{"key":"7475_CR24","doi-asserted-by":"publisher","unstructured":"Mai Z, Li R, Kim H et\u00a0al (2021) Supervised contrastive replay: revisiting the nearest class mean classifier in online class-incremental continual learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 3589\u20133599, https:\/\/doi.org\/10.1109\/cvprw53098.2021.00398","DOI":"10.1109\/cvprw53098.2021.00398"},{"key":"7475_CR25","doi-asserted-by":"crossref","unstructured":"Pryzant R, Yang Z, Xu Y et al. (2022) Automatic rule induction for efficient semi-supervised learning. Empirical Methods in Natural Language Processing https:\/\/doi.org\/10.18653\/v1\/2022.findings-emnlp.3","DOI":"10.18653\/v1\/2022.findings-emnlp.3"},{"key":"7475_CR26","doi-asserted-by":"publisher","unstructured":"Qin C, Joty S (2022) Continual few-shot relation learning via embedding space regularization and data augmentation. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) pp 2776\u20132789. https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.198","DOI":"10.18653\/v1\/2022.acl-long.198"},{"key":"7475_CR27","unstructured":"Qin Q, Hu W, Peng H et\u00a0al (2021) Bns: Building network structures dynamically for continual learning. Advances in Neural Information Processing Systems 34:20608\u201320620. URL https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2021\/file\/ac64504cc249b070772848642cffe6ff-Paper.pdf"},{"key":"7475_CR28","doi-asserted-by":"publisher","unstructured":"Rebuffi SA, Kolesnikov A, Sperl G et\u00a0al (2017) icarl:incremental classifier and representation learning. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp 5533\u20135542, https:\/\/doi.org\/10.1109\/cvpr.2017.587","DOI":"10.1109\/cvpr.2017.587"},{"key":"7475_CR29","doi-asserted-by":"publisher","unstructured":"Rezende D, Danihelka I, Gregor K et\u00a0al (2016) One-shot generalization in deep generative models. In: International Conference on Machine Learning, PMLR, pp 1521\u20131529, https:\/\/doi.org\/10.48550\/arXiv.1603.05106","DOI":"10.48550\/arXiv.1603.05106"},{"key":"7475_CR30","doi-asserted-by":"publisher","unstructured":"Rolnick D, Ahuja A, Schwarz J et\u00a0al (2019) Experience replay for continual learning. Advances in Neural Information Processing Systems 32. https:\/\/doi.org\/10.48550\/arXiv.1811.11682","DOI":"10.48550\/arXiv.1811.11682"},{"key":"7475_CR31","doi-asserted-by":"publisher","first-page":"4439","DOI":"10.1007\/s12652-023-04561-5","volume":"14","author":"Said Karam QMuHLPHLShanq-Jang Ruan","year":"2023","unstructured":"Said Karam QMuHLPHLShanq-Jang Ruan (2023) Episodic memory based continual learning without catastrophic forgetting for environmental sound classification. J Amb Intell Human Comput 14:4439\u20134449. https:\/\/doi.org\/10.1007\/s12652-023-04561-5","journal-title":"J Amb Intell Human Comput"},{"key":"7475_CR32","doi-asserted-by":"publisher","unstructured":"Shi P, Lin J (2019) Simple Bert models for relation extraction and semantic role labeling. arxiv https:\/\/doi.org\/10.48550\/arXiv.1904.05255","DOI":"10.48550\/arXiv.1904.05255"},{"key":"7475_CR33","doi-asserted-by":"publisher","unstructured":"Tran, Thanh, Anh H (2024) Preserving generalization of language models in few-shot continual relation extraction. In: Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pp 13771\u201313784, https:\/\/doi.org\/10.48550\/arXiv.2410.00334","DOI":"10.48550\/arXiv.2410.00334"},{"key":"7475_CR34","doi-asserted-by":"publisher","unstructured":"Van\u00a0de Ven GM, Tolias AS (2019) Three scenarios for continual learning. arXiv preprint arXiv:1904.07734https:\/\/doi.org\/10.48550\/arXiv.1904.07734","DOI":"10.48550\/arXiv.1904.07734"},{"key":"7475_CR35","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2024.3367329","author":"L Wang","year":"2024","unstructured":"Wang L, Zhang X, Su H et al (2024) A comprehensive survey of continual learning: theory, method and application. IEEE Trans Pattern Anal Mach Intell. https:\/\/doi.org\/10.1109\/tpami.2024.3367329","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"7475_CR36","doi-asserted-by":"publisher","unstructured":"Wang X, Wang Z, Hu W (2023) Serial contrastive knowledge distillation for continual few-shot relation extraction. Findings of the Association for Computational Linguistics: ACL 2023 p 12693-12706. https:\/\/doi.org\/10.48550\/arXiv.2305.06616","DOI":"10.48550\/arXiv.2305.06616"},{"key":"7475_CR37","doi-asserted-by":"publisher","unstructured":"Wu Y, Bamman D, Russell S (2017) Adversarial training for relation extraction. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp 1778\u20131783, https:\/\/doi.org\/10.18653\/v1\/d17-1187","DOI":"10.18653\/v1\/d17-1187"},{"key":"7475_CR38","unstructured":"Ye W, Zhang P, Zhang J et\u00a0al (2024) Distilling causal effect of data in continual few-shot relation learning. In: LREC-COLING 2024, p 5041-5051, URL https:\/\/aclanthology.org\/2024.lrec-main.451\/"},{"issue":"2","key":"7475_CR39","doi-asserted-by":"publisher","first-page":"530","DOI":"10.1109\/tkde.2023.3289879","volume":"36","author":"D Zhang","year":"2023","unstructured":"Zhang D, Liu Z, Jia W et al (2023) Dual attention graph convolutional network for relation extraction. IEEE Trans Knowl Data Eng 36(2):530\u2013543. https:\/\/doi.org\/10.1109\/tkde.2023.3289879","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"7475_CR40","doi-asserted-by":"publisher","unstructured":"Zhang Y, Zhong V, Chen D et\u00a0al (2017) Position-aware attention and supervised data improve slot filling. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp 35\u201345, https:\/\/doi.org\/10.18653\/v1\/D17-1004","DOI":"10.18653\/v1\/D17-1004"},{"issue":"8","key":"7475_CR41","doi-asserted-by":"publisher","first-page":"7953","DOI":"10.1109\/tkde.2022.3161584","volume":"35","author":"X Zhao","year":"2022","unstructured":"Zhao X, Yang M, Qu Q et al (2022) Exploring privileged features for relation extraction with contrastive student-teacher learning. IEEE Trans Knowl Data Eng 35(8):7953\u20137965. https:\/\/doi.org\/10.1109\/tkde.2022.3161584","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"7475_CR42","doi-asserted-by":"publisher","unstructured":"Zhou J, Zheng Y, Tang J et\u00a0al (2021) Flipda: effective and robust data augmentation for few-shot learning. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) pp 8646-8665. https:\/\/doi.org\/10.48550\/arXiv.2108.06332","DOI":"10.48550\/arXiv.2108.06332"}],"container-title":["The Journal of Supercomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-07475-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11227-025-07475-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-07475-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T13:33:19Z","timestamp":1749130399000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11227-025-07475-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,5]]},"references-count":42,"journal-issue":{"issue":"8","published-online":{"date-parts":[[2025,6]]}},"alternative-id":["7475"],"URL":"https:\/\/doi.org\/10.1007\/s11227-025-07475-5","relation":{},"ISSN":["1573-0484"],"issn-type":[{"value":"1573-0484","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,5]]},"assertion":[{"value":"16 May 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 June 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"969"}}