{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,21]],"date-time":"2025-06-21T04:30:37Z","timestamp":1750480237506,"version":"3.32.0"},"reference-count":8,"publisher":"Association for Natural Language Processing","issue":"4","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Journal of Natural Language Processing"],"published-print":{"date-parts":[[2024]]},"DOI":"10.5715\/jnlp.31.1767","type":"journal-article","created":{"date-parts":[[2024,12,14]],"date-time":"2024-12-14T22:14:29Z","timestamp":1734214469000},"page":"1767-1773","source":"Crossref","is-referenced-by-count":2,"title":["Does Pre-trained Language Model Actually Infer Unseen Links in Knowledge Graph Completion?"],"prefix":"10.5715","volume":"31","author":[{"given":"Yusuke","family":"Sakai","sequence":"first","affiliation":[{"name":"Nara Institute of Science and Technology"}]}],"member":"3685","reference":[{"unstructured":"Bordes, A., Usunier, N., Garcia-Dur\u00e1n, A., Weston, J., and Yakhnenko, O. (2013). \u201cTranslating Embeddings for Modeling Multi-Relational Data.\u201d In <i>Proceedings of the 26th International Conference on Neural Information Processing Systems - Volume 2, NIPS\u201913<\/i>, pp. 2787\u20132795, Red Hook, NY, USA. Curran Associates Inc.","key":"1"},{"unstructured":"Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). \u201cBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.\u201d In Burstein, J., Doran, C., and Solorio, T. (Eds.), <i>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)<\/i>, pp. 4171\u20134186, Minneapolis, Minnesota. Association for Computational Linguistics.","key":"2"},{"unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., and Liu, P. J. (2020). \u201cExploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.\u201d <i>Journal of Machine Learning Research<\/i>, 21 (1), pp. 5485\u20135551.","key":"3"},{"doi-asserted-by":"crossref","unstructured":"Sakai, Y., Kamigaito, H., Hayashi, K., and Watanabe, T. (2024). \u201cDoes Pre-trained Language Model Actually Infer Unseen Links in Knowledge Graph Completion?\u201d In Duh, K., Gomez, H., and Bethard, S. (Eds.), <i>Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)<\/i>, pp. 8091\u20138106, Mexico City, Mexico. Association for Computational Linguistics.","key":"4","DOI":"10.18653\/v1\/2024.naacl-long.447"},{"doi-asserted-by":"crossref","unstructured":"Saxena, A., Kochsiek, A., and Gemulla, R. (2022). \u201cSequence-to-Sequence Knowledge Graph Completion and Question Answering.\u201d In Muresan, S., Nakov, P., and Villavicencio, A. (Eds.), <i>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)<\/i>, pp. 2814\u20132828, Dublin, Ireland. Association for Computational Linguistics.","key":"5","DOI":"10.18653\/v1\/2022.acl-long.201"},{"unstructured":"Trouillon, T., Welbl, J., Riedel, S., Gaussier, \u00c9., and Bouchard, G. (2016). \u201cComplex Embeddings for Simple Link Prediction.\u201d In <i>Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48, ICML\u201916<\/i>, pp. 2071\u20132080. JMLR.org.","key":"6"},{"doi-asserted-by":"crossref","unstructured":"Wang, L., Zhao, W., Wei, Z., and Liu, J. (2022). \u201cSimKGC: Simple Contrastive Knowledge Graph Completion with Pre-trained Language Models.\u201d In Muresan, S., Nakov, P., and Villavicencio, A. (Eds.), <i>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)<\/i>, pp. 4281\u20134294, Dublin, Ireland. Association for Computational Linguistics.","key":"7","DOI":"10.18653\/v1\/2022.acl-long.295"},{"doi-asserted-by":"crossref","unstructured":"Wang, X., Gao, T., Zhu, Z., Zhang, Z., Liu, Z., Li, J., and Tang, J. (2021). \u201cKEPLER: A Unified Model for Knowledge Embedding and Pre-trained Language Representation.\u201d <i>Transactions of the Association for Computational Linguistics<\/i>, 9, pp. 176\u2013194.","key":"8","DOI":"10.1162\/tacl_a_00360"}],"container-title":["Journal of Natural Language Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/jnlp\/31\/4\/31_1767\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,21]],"date-time":"2024-12-21T04:31:52Z","timestamp":1734755512000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/jnlp\/31\/4\/31_1767\/_article\/-char\/ja\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":8,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024]]}},"URL":"https:\/\/doi.org\/10.5715\/jnlp.31.1767","relation":{},"ISSN":["1340-7619","2185-8314"],"issn-type":[{"type":"print","value":"1340-7619"},{"type":"electronic","value":"2185-8314"}],"subject":[],"published":{"date-parts":[[2024]]}}}