{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,6]],"date-time":"2025-08-06T12:33:47Z","timestamp":1754483627154,"version":"3.37.3"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T00:00:00Z","timestamp":1721606400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T00:00:00Z","timestamp":1721606400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Department of Education of Hunan province","award":["21B0462"],"award-info":[{"award-number":["21B0462"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U2003208"],"award-info":[{"award-number":["U2003208"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Prompt-tuning has emerged as a promising approach for improving the performance of classification tasks by converting them into masked language modeling problems through the insertion of text templates. Despite its considerable success, applying this approach to relation extraction is challenging. Predicting the relation, often expressed as a specific word or phrase between two entities, usually requires creating mappings from these terms to an existing lexicon and introducing extra learnable parameters. This can lead to a decrease in coherence between the pre-training task and fine-tuning. To address this issue, we propose a novel method for prompt-tuning in relation extraction, aiming to enhance the coherence between fine-tuning and pre-training tasks. Specifically, we avoid the need for a suitable relation word by converting the relation into relational semantic keywords, which are representative phrases that encapsulate the essence of the relation. Moreover, we employ a composite loss function that optimizes the model at both token and relation levels. Our approach incorporates the masked language modeling (MLM) loss and the entity pair constraint loss for predicted tokens. For relation level optimization, we use both the cross-entropy loss and TransE. Extensive experimental results on four datasets demonstrate that our method significantly improves performance in relation extraction tasks. The results show an average improvement of approximately 1.6 points in <jats:italic>F<\/jats:italic>1 metrics compared to the current state-of-the-art model. Codes are released at <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/12138yx\/TCohPrompt\">https:\/\/github.com\/12138yx\/TCohPrompt<\/jats:ext-link>.<\/jats:p>","DOI":"10.1007\/s40747-024-01563-4","type":"journal-article","created":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T06:01:50Z","timestamp":1721628110000},"page":"7565-7575","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["TCohPrompt: task-coherent prompt-oriented fine-tuning for relation extraction"],"prefix":"10.1007","volume":"10","author":[{"given":"Jun","family":"Long","sequence":"first","affiliation":[]},{"given":"Zhuoying","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Chao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Wenti","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,22]]},"reference":[{"key":"1563_CR1","doi-asserted-by":"crossref","unstructured":"Alt C, Gabryszak A, Hennig L (2020) TACRED revisited: a thorough evaluation of the TACRED relation extraction task. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 1558\u20131569","DOI":"10.18653\/v1\/2020.acl-main.142"},{"key":"1563_CR2","first-page":"2787","volume":"26","author":"A Bordes","year":"2013","unstructured":"Bordes A, Usunier N, Garcia-Duran A et al (2013) Translating embeddings for modeling multi-relational data. Adv Neural Inf Process Syst 26:2787\u20132795","journal-title":"Adv Neural Inf Process Syst"},{"key":"1563_CR3","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"key":"1563_CR4","doi-asserted-by":"crossref","unstructured":"Chen X, Li L, Zhang N et al (2022) Relation extraction as open-book examination: retrieval-enhanced prompt tuning. In: Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pp 2443\u20132448","DOI":"10.1145\/3477495.3531746"},{"key":"1563_CR5","first-page":"2778","volume":"2022","author":"X Chen","year":"2022","unstructured":"Chen X, Zhang N, Xie X et al (2022) KnowPrompt: knowledge-aware prompt-tuning with synergistic optimization for relation extraction. Proc ACM Web Conf 2022:2778\u20132788","journal-title":"Proc ACM Web Conf"},{"key":"1563_CR6","doi-asserted-by":"crossref","unstructured":"Culotta A, Sorensen J (2004) Dependency tree kernels for relation extraction. In: Proceedings of the 42nd annual meeting of the association for computational linguistics (ACL-04), pp 423\u2013429","DOI":"10.3115\/1218955.1219009"},{"key":"1563_CR7","doi-asserted-by":"publisher","unstructured":"Devlin J, Chang MW, Lee K et al (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers). Association for Computational Linguistics, Minneapolis, Minnesota, pp 4171\u20134186. https:\/\/doi.org\/10.18653\/v1\/N19-1423. https:\/\/aclanthology.org\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"1563_CR8","doi-asserted-by":"publisher","unstructured":"Gao T, Fisch A, Chen D (2021) Making pre-trained language models better few-shot learners. In: Zong C, Xia F, Li W et\u00a0al (eds) Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: long papers). Association for Computational Linguistics, Online, pp 3816\u20133830. https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.295. https:\/\/aclanthology.org\/2021.acl-long.295","DOI":"10.18653\/v1\/2021.acl-long.295"},{"key":"1563_CR9","unstructured":"Guti\u00e9rrez BJ, McNeal N, Washington C et al (2022) Thinking about GPT-3 in-context learning for biomedical IE? Think again. arXiv:2203.08410"},{"key":"1563_CR10","doi-asserted-by":"crossref","unstructured":"Han X, Liu Z, Sun M (2018) Neural knowledge acquisition via mutual attention between knowledge graph and text. In: Proceedings of the AAAI conference on artificial intelligence, pp 4832\u20134839","DOI":"10.1609\/aaai.v32i1.11927"},{"key":"1563_CR11","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1016\/j.aiopen.2022.11.003","volume":"3","author":"X Han","year":"2022","unstructured":"Han X, Zhao W, Ding N et al (2022) PTR: prompt tuning with rules for text classification. AI Open 3:182\u2013192","journal-title":"AI Open"},{"key":"1563_CR12","doi-asserted-by":"crossref","unstructured":"Hendrickx I, Kim SN, Kozareva Z et\u00a0al (2010) SemEval-2010 task 8: multi-way classification of semantic relations between pairs of nominals. In: Proceedings of the 5th international workshop on semantic evaluation. Association for Computational Linguistics, Uppsala, Sweden, pp 33\u201338. https:\/\/aclanthology.org\/S10-1006","DOI":"10.3115\/1621969.1621986"},{"key":"1563_CR13","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1162\/tacl_a_00300","volume":"8","author":"M Joshi","year":"2020","unstructured":"Joshi M, Chen D, Liu Y et al (2020) SpanBERT: improving pre-training by representing and predicting spans. Trans Assoc Comput Linguist 8:64\u201377","journal-title":"Trans Assoc Comput Linguist"},{"key":"1563_CR14","unstructured":"Kojima T, Gu SS, Reid M et al (2023) Large language models are zero-shot reasoners. arXiv:2205.11916"},{"key":"1563_CR15","doi-asserted-by":"publisher","unstructured":"Lewis M, Liu Y, Goyal N et\u00a0al (2020) BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, pp 7871\u20137880. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.703. https:\/\/aclanthology.org\/2020.acl-main.703","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"1563_CR16","doi-asserted-by":"crossref","unstructured":"Lin Y, Shen S, Liu Z et al (2016) Neural relation extraction with selective attention over instances. In: Proceedings of the 54th annual meeting of the association for computational linguistics (volume 1: long papers), pp 2124\u20132133","DOI":"10.18653\/v1\/P16-1200"},{"issue":"9","key":"1563_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu P, Yuan W, Fu J et al (2023) Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput Surv 55(9):1\u201335","journal-title":"ACM Comput Surv"},{"key":"1563_CR18","doi-asserted-by":"crossref","unstructured":"Liu T, Zhang X, Zhou W et al (2018) Neural relation extraction via inner-sentence noise reduction and transfer learning. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 2195\u20132204","DOI":"10.18653\/v1\/D18-1243"},{"key":"1563_CR19","unstructured":"Liu Y, Ott M, Goyal N et\u00a0al (2019) RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692"},{"key":"1563_CR20","unstructured":"Minard AL, Ligozat AL, Grau B (2011) Multi-class SVM for relation extraction from clinical reports. In: Proceedings of the international conference recent advances in natural language processing, 2011, pp 604\u2013609"},{"key":"1563_CR21","doi-asserted-by":"crossref","unstructured":"Mintz M, Bills S, Snow R et al (2009) Distant supervision for relation extraction without labeled data. In: Proceedings of the joint conference of the 47th annual meeting of the ACL and the 4th international joint conference on natural language processing of the AFNLP, pp 1003\u20131011","DOI":"10.3115\/1690219.1690287"},{"key":"1563_CR22","doi-asserted-by":"crossref","unstructured":"Peters ME, Neumann M, Logan R et al (2019) Knowledge enhanced contextual word representations. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 43\u201354","DOI":"10.18653\/v1\/D19-1005"},{"key":"1563_CR23","doi-asserted-by":"crossref","unstructured":"Petroni F, Rockt\u00e4schel T, Riedel S et\u00a0al (2019) Language models as knowledge bases? In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 2463\u20132473","DOI":"10.18653\/v1\/D19-1250"},{"key":"1563_CR24","unstructured":"Radford A, Narasimhan K, Salimans T et\u00a0al (2018) Improving language understanding by generative pre-training"},{"key":"1563_CR25","doi-asserted-by":"crossref","unstructured":"Sainz O, de\u00a0Lacalle OL, Labaka G et\u00a0al (2021) Label verbalization and entailment for effective zero and few-shot relation extraction. In: Proceedings of the 2021 conference on empirical methods in natural language processing, pp 1199\u20131212","DOI":"10.18653\/v1\/2021.emnlp-main.92"},{"key":"1563_CR26","doi-asserted-by":"crossref","unstructured":"Schick T, Sch\u00fctze H (2021) Exploiting cloze-questions for few-shot text classification and natural language inference. In: Proceedings of the 16th conference of the European chapter of the association for computational linguistics: main volume, pp 255\u2013269","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"1563_CR27","doi-asserted-by":"publisher","unstructured":"Schick T, Sch\u00fctze H (2021) Exploiting cloze-questions for few-shot text classification and natural language inference. In: Merlo P, Tiedemann J, Tsarfaty R (eds) Proceedings of the 16th conference of the european chapter of the association for computational linguistics: main volume. Association for Computational Linguistics, Online, pp 255\u2013269. https:\/\/doi.org\/10.18653\/v1\/2021.eacl-main.20. https:\/\/aclanthology.org\/2021.eacl-main.20","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"1563_CR28","doi-asserted-by":"publisher","unstructured":"Schick T, Schmid H, Sch\u00fctze H (2020) Automatically identifying words that can serve as labels for few-shot text classification. In: Scott D, Bel N, Zong C (eds) Proceedings of the 28th international conference on computational linguistics. international committee on computational linguistics, Barcelona, Spain (Online), pp 5569\u20135578. https:\/\/doi.org\/10.18653\/v1\/2020.coling-main.488. https:\/\/aclanthology.org\/2020.coling-main.488","DOI":"10.18653\/v1\/2020.coling-main.488"},{"key":"1563_CR29","doi-asserted-by":"crossref","unstructured":"Soares LB, Fitzgerald N, Ling J et\u00a0al (2019) Matching the blanks: distributional similarity for relation learning. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 2895\u20132905","DOI":"10.18653\/v1\/P19-1279"},{"key":"1563_CR30","doi-asserted-by":"crossref","unstructured":"Stoica G, Platanios EA, P\u00f3czos B (2021) Re-TACRED: addressing shortcomings of the TACRED dataset. In: Proceedings of the AAAI conference on artificial intelligence, pp 13843\u201313850","DOI":"10.1609\/aaai.v35i15.17631"},{"key":"1563_CR31","doi-asserted-by":"crossref","unstructured":"Wadhwa S, Amir S, Wallace BC (2023) Revisiting relation extraction in the era of large language models. arXiv preprint arXiv:2305.05003","DOI":"10.18653\/v1\/2023.acl-long.868"},{"key":"1563_CR32","doi-asserted-by":"publisher","unstructured":"Wan Z, Cheng F, Mao Z et al (2023) GPT-RE: in-context learning for relation extraction using large language models. In: Bouamor H, Pino J, Bali K (eds) Proceedings of the 2023 conference on empirical methods in natural language processing. Association for Computational Linguistics, Singapore, pp 3534\u20133547. https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.214. https:\/\/aclanthology.org\/2023.emnlp-main.214","DOI":"10.18653\/v1\/2023.emnlp-main.214"},{"key":"1563_CR33","doi-asserted-by":"crossref","unstructured":"Wan Z, Cheng F, Mao Z et\u00a0al (2023b) GPT-RE: in-context learning for relation extraction using large language models. arXiv preprint arXiv:2305.02105","DOI":"10.18653\/v1\/2023.emnlp-main.214"},{"key":"1563_CR34","unstructured":"Wang X, Wei J, Schuurmans D et\u00a0al (2023) Self-consistency improves chain of thought reasoning in language models. arXiv:2203.11171"},{"key":"1563_CR35","doi-asserted-by":"crossref","unstructured":"Wu S, He Y (2019) Enriching pre-trained language model with entity information for relation classification. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp 2361\u20132364","DOI":"10.1145\/3357384.3358119"},{"key":"1563_CR36","doi-asserted-by":"crossref","unstructured":"Xue F, Sun A, Zhang H et al (2021) GDPNet: refining latent multi-view graph for relation extraction. In: Proceedings of the AAAI conference on artificial intelligence, pp 14194\u201314202","DOI":"10.1609\/aaai.v35i16.17670"},{"key":"1563_CR37","doi-asserted-by":"crossref","unstructured":"Yamada I, Asai A, Shindo H et al (2020) LUKE: deep contextualized entity representations with entity-aware self-attention. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 6442\u20136454","DOI":"10.18653\/v1\/2020.emnlp-main.523"},{"key":"1563_CR38","unstructured":"Zeng D, Liu K, Lai S et al (2014) Relation classification via convolutional deep neural network. In: Proceedings of COLING 2014, the 25th international conference on computational linguistics: technical papers, pp 2335\u20132344"},{"key":"1563_CR39","doi-asserted-by":"crossref","unstructured":"Zhang P, Lan C, Zeng W et al (2020) Semantics-guided neural networks for efficient skeleton-based human action recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1112\u20131121","DOI":"10.1109\/CVPR42600.2020.00119"},{"key":"1563_CR40","doi-asserted-by":"crossref","unstructured":"Zhang Y, Zhong V, Chen D et al (2017) Position-aware attention and supervised data improve slot filling. In: Proceedings of the 2017 conference on empirical methods in natural language processing, pp 35\u201345","DOI":"10.18653\/v1\/D17-1004"},{"key":"1563_CR41","doi-asserted-by":"crossref","unstructured":"Zhou G, Su J, Zhang J et al (2005) Exploring various knowledge in relation extraction. In: Proceedings of the 43rd annual meeting of the association for computational linguistics (ACL\u201905), pp 427\u2013434","DOI":"10.3115\/1219840.1219893"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01563-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01563-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01563-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,16]],"date-time":"2024-10-16T22:08:10Z","timestamp":1729116490000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01563-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,22]]},"references-count":41,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["1563"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01563-4","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"type":"print","value":"2199-4536"},{"type":"electronic","value":"2198-6053"}],"subject":[],"published":{"date-parts":[[2024,7,22]]},"assertion":[{"value":"26 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors confirm that they have no affiliations or involvement with any organization or entity that has a financial or non-financial interest in the subject matter or materials discussed in this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}