{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T02:51:36Z","timestamp":1768877496176,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":55,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819712762","type":"print"},{"value":"9789819712779","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-1277-9_6","type":"book-chapter","created":{"date-parts":[[2024,4,2]],"date-time":"2024-04-02T02:01:41Z","timestamp":1712023301000},"page":"73-84","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["A Review of Relationship Extraction Based on Deep Learning"],"prefix":"10.1007","author":[{"given":"Guolong","family":"Liao","sequence":"first","affiliation":[]},{"given":"Xiangyan","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Tian","family":"Li","sequence":"additional","affiliation":[]},{"given":"Li","family":"Zhong","sequence":"additional","affiliation":[]},{"given":"Pengfan","family":"Zeng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,3]]},"reference":[{"key":"6_CR1","doi-asserted-by":"crossref","unstructured":"Zeng, W., Lin, Y., Liu, Z., et al.: Incorporating relation paths in neural relation extraction. arXiv preprint arXiv:1609.07479 (2016)","DOI":"10.18653\/v1\/D17-1186"},{"key":"6_CR2","doi-asserted-by":"crossref","unstructured":"Liu, G., Fu, L., Yu, B., Cui, L.: Automatic recognition of parallel sentence based on sentences-interaction CNN and its application. In: 2022 7th International Conference on Computer and Communication Systems (ICCCS), pp. 245\u2013250. IEEE, April 2022","DOI":"10.1109\/ICCCS55155.2022.9846217"},{"issue":"6","key":"6_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2020.102373","volume":"57","author":"H Wen","year":"2020","unstructured":"Wen, H., Zhu, X., Zhang, L., Li, F.: A gated piecewise CNN with entity-aware enhancement for distantly supervised relation extraction. Inf. Process. Manage. 57(6), 102373 (2020)","journal-title":"Inf. Process. Manage."},{"key":"6_CR4","doi-asserted-by":"crossref","unstructured":"Li, Y., Ni, P., Li, G., Chang, V.: Effective piecewise CNN with attention mechanism for distant supervision on relation extraction task. In: 5th International Conference on Complexity, Future Information Systems and Risk, SciTePress. pp. 53\u201362, May 2020","DOI":"10.5220\/0009582700530060"},{"key":"6_CR5","unstructured":"Liu, Z., Huang, H., Lu, C., Lyu, S.: Multichannel CNN with attention for text classification. arXiv preprint arXiv:2006.16174 (2020)"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Shimura, K., Li, J., Fukumoto, F.: HFT-CNN: learning hierarchical category structure for multi-label short text categorization. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 811\u2013816 (2018)","DOI":"10.18653\/v1\/D18-1093"},{"key":"6_CR7","unstructured":"Chung, J., Gulcehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 (2014)"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Smirnova, E., Vasile, F.: Contextual sequence modeling for recommendation with recurrent neural networks. In: Proceedings of the 2nd Workshop on Deep Learning for Recommender Systems, pp. 2\u20139, August 2017","DOI":"10.1145\/3125486.3125488"},{"issue":"1","key":"6_CR9","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1109\/TKDE.2020.2981333","volume":"34","author":"Z Zhang","year":"2020","unstructured":"Zhang, Z., Cui, P., Zhu, W.: Deep learning on graphs: a survey. IEEE Trans. Knowl. Data Eng. 34(1), 249\u2013270 (2020)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"6_CR10","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"6_CR11","unstructured":"Kenton, J.D.M.W.C., Toutanova, L.K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of naacL-HLT, vol. 1, p. 2, June 2019"},{"key":"6_CR12","unstructured":"Zeng, D., Liu, K., Lai, S., Zhou, G., Zhao, J.: Relation classification via convolutional deep neural network. In: Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, pp. 2335\u20132344, August 2014"},{"key":"6_CR13","doi-asserted-by":"publisher","first-page":"279","DOI":"10.1016\/j.future.2020.08.005","volume":"115","author":"ME Basiri","year":"2021","unstructured":"Basiri, M.E., Nemati, S., Abdar, M., Cambria, E., Acharya, U.R.: ABCDM: an attention-based bidirectional CNN-RNN deep model for sentiment analysis. Futur. Gener. Comput. Syst. 115, 279\u2013294 (2021)","journal-title":"Futur. Gener. Comput. Syst."},{"key":"6_CR14","doi-asserted-by":"publisher","first-page":"141801","DOI":"10.1109\/ACCESS.2019.2943545","volume":"7","author":"K Sun","year":"2019","unstructured":"Sun, K., Li, Y., Deng, D., Li, Y.: Multi-channel CNN based inner-attention for compound sentence relation classification. IEEE Access 7, 141801\u2013141809 (2019)","journal-title":"IEEE Access"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Nguyen, T.H., Grishman, R.: Relation extraction: perspective from convolutional neural networks. In: Proceedings of the 1st Workshop on Vector Space Modeling for Natural Language Processing, pp. 39\u201348, June 2015","DOI":"10.3115\/v1\/W15-1506"},{"key":"6_CR16","unstructured":"Santos, C.N.D., Xiang, B., Zhou, B.: Classifying relations by ranking with convolutional neural networks. arXiv preprint  arXiv:1504.06580 (2015)"},{"key":"6_CR17","unstructured":"Zhang, D., Wang, D.: Relation classification via recurrent neural network. arXiv preprint arXiv:1508.01006 (2015)"},{"key":"6_CR18","unstructured":"Zhang, S., Zheng, D., Hu, X., Yang, M.: Bidirectional long short-term memory networks for relation classification. In: Proceedings of the 29th Pacific Asia Conference on Language, Information and Computation, pp. 73\u201378, October 2015"},{"key":"6_CR19","unstructured":"Xu, Y., et al.: Improved relation classification by deep recurrent neural networks with data augmentation. arXiv preprint arXiv:1601.03651 (2016)"},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Shen, Y., Ma, X., Tang, Y., Lu, W.: A trigger-sense memory flow framework for joint entity and relation extraction. In: Proceedings of the Web Conference 2021, pp. 1704\u20131715, April 2021","DOI":"10.1145\/3442381.3449895"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Zhong, Z., Chen, D.: A frustratingly easy approach for entity and relation extraction. arXiv preprint arXiv:2010.12812 (2020)","DOI":"10.18653\/v1\/2021.naacl-main.5"},{"key":"6_CR22","doi-asserted-by":"crossref","unstructured":"Wang, J., Lu, W.: Two are better than one: Joint entity and relation extraction with table-sequence encoders. arXiv preprint  arXiv:2010.03851 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.133"},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Wang, Y., Sun, C., Wu, Y., Zhou, H., Li, L., Yan, J.: UniRE: a unified label space for entity relation extraction. arXiv preprint arXiv:2107.04292 (2021)","DOI":"10.18653\/v1\/2021.acl-long.19"},{"key":"6_CR24","doi-asserted-by":"crossref","unstructured":"Ye, D., Lin, Y., Li, P., Sun, M.: Packed levitated marker for entity and relation extraction. arXiv preprint arXiv:2109.06067 (2021)","DOI":"10.18653\/v1\/2022.acl-long.337"},{"key":"6_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-021-00492-0","volume":"8","author":"C Shorten","year":"2021","unstructured":"Shorten, C., Khoshgoftaar, T.M., Furht, B.: Text data augmentation for deep learning. J. big Data 8, 1\u201334 (2021)","journal-title":"J. big Data"},{"issue":"7","key":"6_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3544558","volume":"55","author":"M Bayer","year":"2022","unstructured":"Bayer, M., Kaufhold, M.A., Reuter, C.: A survey on data augmentation for text classification. ACM Comput. Surv. 55(7), 1\u201339 (2022)","journal-title":"ACM Comput. Surv."},{"key":"6_CR27","doi-asserted-by":"crossref","unstructured":"Liu, P., Wang, X., Xiang, C., Meng, W.: A survey of text data augmentation. In: 2020 International Conference on Computer Communication and Network Security (CCNS), pp. 191\u2013195. IEEE, August 2020","DOI":"10.1109\/CCNS50731.2020.00049"},{"key":"6_CR28","doi-asserted-by":"crossref","unstructured":"Wei, J., Zou, K.: EDA: easy data augmentation techniques for boosting performance on text classification tasks. arXiv preprint arXiv:1901.11196 (2019)","DOI":"10.18653\/v1\/D19-1670"},{"key":"6_CR29","doi-asserted-by":"crossref","unstructured":"Karimi, A., Rossi, L., Prati, A.: AEDA: an easier data augmentation technique for text classification. arXiv preprint arXiv:2108.13230 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.234"},{"key":"6_CR30","doi-asserted-by":"crossref","unstructured":"Rizos, G., Hemker, K., Schuller, B.: Augment to prevent: short-text data augmentation in deep learning for hate-speech classification. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 991\u20131000, November 2019","DOI":"10.1145\/3357384.3358040"},{"key":"6_CR31","unstructured":"Papanikolaou, Y., Pierleoni, A.: Dare: Data augmented relation extraction with gpt-2. arXiv preprint arXiv:2004.13845 (2020)"},{"key":"6_CR32","doi-asserted-by":"crossref","unstructured":"Wang, A., et al.: Entity relation extraction in the medical domain: based on data augmentation. Ann. Trans. Med. 10(19) (2022)","DOI":"10.21037\/atm-22-3991"},{"key":"6_CR33","doi-asserted-by":"crossref","unstructured":"Hu, X.: GDA: Generative Data Augmentation Techniques for Relation Extraction Tasks. arXiv preprint arXiv:2305.16663 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.649"},{"issue":"7","key":"6_CR34","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0216913","volume":"14","author":"P Su","year":"2019","unstructured":"Su, P., Li, G., Wu, C., Vijay-Shanker, K.: Using distant supervision to augment manually annotated data for relation extraction. PLoS ONE 14(7), e0216913 (2019)","journal-title":"PLoS ONE"},{"key":"6_CR35","doi-asserted-by":"crossref","unstructured":"Qin, P., Xu, W., Wang, W.Y.: Robust distant supervision relation extraction via deep reinforcement learning. arXiv preprint arXiv:1805.09927 (2018)","DOI":"10.18653\/v1\/P18-1199"},{"key":"6_CR36","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1016\/j.neunet.2021.04.032","volume":"142","author":"Y Zhou","year":"2021","unstructured":"Zhou, Y., Pan, L., Bai, C., Luo, S., Wu, Z.: Self-selective attention using correlation between instances for distant supervision relation extraction. Neural Netw. 142, 213\u2013220 (2021)","journal-title":"Neural Netw."},{"issue":"5","key":"6_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3241741","volume":"51","author":"A Smirnova","year":"2018","unstructured":"Smirnova, A., Cudr\u00e9-Mauroux, P.: Relation extraction using distant supervision: a survey. ACM Comput. Surv. (CSUR) 51(5), 1\u201335 (2018)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"6_CR38","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1016\/j.ins.2019.09.006","volume":"509","author":"Z Geng","year":"2020","unstructured":"Geng, Z., Chen, G., Han, Y., Lu, G., Li, F.: Semantic relation extraction using sequential and tree-structured LSTM with attention. Inf. Sci. 509, 183\u2013192 (2020)","journal-title":"Inf. Sci."},{"key":"6_CR39","doi-asserted-by":"publisher","unstructured":"Yang, T. et al.: Tree-capsule: tree-structured capsule network for improving relation extraction. In: Karlapalem, K., et al. Advances in Knowledge Discovery and Data Mining. PAKDD 2021. Lecture Notes in Computer Science(), vol 12714. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-75768-7_26","DOI":"10.1007\/978-3-030-75768-7_26"},{"key":"6_CR40","unstructured":"Shen, Y., Huang, X.J.: Attention-based convolutional neural network for semantic relation extraction. In: Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pp. 2526\u20132536, December 2016"},{"key":"6_CR41","doi-asserted-by":"crossref","unstructured":"Yuan, Y., Zhou, X., Pan, S., Zhu, Q., Song, Z., Guo, L.: A relation-specific attention network for joint entity and relation extraction. In: International Joint Conference on Artificial Intelligence, January 2021","DOI":"10.24963\/ijcai.2020\/561"},{"key":"6_CR42","doi-asserted-by":"crossref","unstructured":"Li, X., et al.: Entity-relation extraction as multi-turn question answering. arXiv preprint arXiv:1905.05529 (2019)","DOI":"10.18653\/v1\/P19-1129"},{"key":"6_CR43","doi-asserted-by":"crossref","unstructured":"Liu, T., Wang, K., Chang, B., Sui, Z.: A soft-label method for noise-tolerant distantly supervised relation extraction. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 1790\u20131795, September 2017","DOI":"10.18653\/v1\/D17-1189"},{"key":"6_CR44","doi-asserted-by":"crossref","unstructured":"Zhang, N., Deng, S., Sun, Z., Chen, X., Zhang, W., Chen, H.: Attention-based capsule networks with dynamic routing for relation extraction. arXiv preprint arXiv:1812.11321 (2018)","DOI":"10.18653\/v1\/D18-1120"},{"key":"6_CR45","doi-asserted-by":"crossref","unstructured":"Alt, C., Gabryszak, A., Hennig, L.: TACRED revisited: A thorough evaluation of the TACRED relation extraction task. arXiv preprint arXiv:2004.14855 (2020)","DOI":"10.18653\/v1\/2020.acl-main.142"},{"key":"6_CR46","unstructured":"Lample, G., Conneau, A.: Cross-lingual language model pretraining. arXiv preprint arXiv:1901.07291 (2019)"},{"issue":"4","key":"6_CR47","doi-asserted-by":"publisher","first-page":"459","DOI":"10.1080\/14790718.2019.1569017","volume":"16","author":"M Poeste","year":"2019","unstructured":"Poeste, M., M\u00fcller, N., Arnaus Gil, L.: Code-mixing and language dominance: bilingual, trilingual and multilingual children compared. Int. J. Multiling. 16(4), 459\u2013491 (2019)","journal-title":"Int. J. Multiling."},{"key":"6_CR48","doi-asserted-by":"crossref","unstructured":"Yao, Y.: CodRED: A cross-document relation extraction dataset for acquiring knowledge in the wild. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 4452\u20134472, November 2021","DOI":"10.18653\/v1\/2021.emnlp-main.366"},{"key":"6_CR49","doi-asserted-by":"crossref","unstructured":"Wang, F., et al.: Entity-centered cross-document relation extraction. arXiv preprint arXiv:2210.16541 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.671"},{"issue":"7","key":"6_CR50","doi-asserted-by":"publisher","first-page":"1608","DOI":"10.1109\/JAS.2023.123540","volume":"10","author":"H Wu","year":"2023","unstructured":"Wu, H., Chen, X., Hu, Z., Shi, J., Xu, S., Xu, B.: Local-to-global causal reasoning for cross-document relation extraction. IEEE\/CAA J. Autom. Sinica 10(7), 1608\u20131621 (2023)","journal-title":"IEEE\/CAA J. Autom. Sinica"},{"key":"6_CR51","unstructured":"Vaswani, A., et al.: Attention is all you need. Advances in neural information processing systems, 30 (2022)"},{"key":"6_CR52","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)"},{"issue":"8","key":"6_CR53","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI blog 1(8), 9 (2019)","journal-title":"OpenAI blog"},{"key":"6_CR54","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"6_CR55","unstructured":"Bubeck, S., et al.: Sparks of artificial general intelligence: early experiments with gpt-4. arXiv preprint arXiv:2303.12712 (2023)"}],"container-title":["Communications in Computer and Information Science","Artificial Intelligence and Machine Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-1277-9_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,23]],"date-time":"2025-10-23T10:44:11Z","timestamp":1761216251000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-1277-9_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819712762","9789819712779"],"references-count":55,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-1277-9_6","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"3 April 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IAIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Artificial Intelligence Conference","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Nanjing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iaic2023a","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.iaicconf.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}