{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T12:01:13Z","timestamp":1743076873304,"version":"3.40.3"},"publisher-location":"Cham","reference-count":42,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031826962"},{"type":"electronic","value":"9783031826979"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-82697-9_14","type":"book-chapter","created":{"date-parts":[[2025,2,15]],"date-time":"2025-02-15T10:17:28Z","timestamp":1739614648000},"page":"184-197","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Knowledge Neurons in\u00a0the\u00a0Knowledge Graph-based Link Prediction Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6738-6970","authenticated-orcid":false,"given":"Grzegorz P.","family":"Mika","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4890-9005","authenticated-orcid":false,"given":"Amel","family":"Bouzeghoub","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9776-3842","authenticated-orcid":false,"given":"Katarzyna","family":"W\u0119grzyn-Wolska","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8924-0769","authenticated-orcid":false,"given":"Yessin M.","family":"Neggaz","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,16]]},"reference":[{"issue":"3","key":"14_CR1","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1016\/S0378-8733(03)00009-1","volume":"25","author":"LA Adamic","year":"2003","unstructured":"Adamic, L.A., Adar, E.: Friends and neighbors on the web. Soc. Netw. 25(3), 211\u2013230 (2003)","journal-title":"Soc. Netw."},{"key":"14_CR2","unstructured":"Bordes, A., Usunier, N., Garcia-Duran, A., Weston, J., Yakhnenko, O.: Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 26 (2013)"},{"key":"14_CR3","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"14_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision \u2013 ECCV 2020","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"14_CR5","doi-asserted-by":"crossref","unstructured":"Chen, J., Gao, K., Li, G., He, K.: NAGphormer: a tokenized graph transformer for node classification in large graphs. In: International Conference on Learning Representations (2023)","DOI":"10.1109\/TBDATA.2024.3524081"},{"key":"14_CR6","doi-asserted-by":"crossref","unstructured":"Chen, Y., Cao, P., Chen, Y., Liu, K., Zhao, J.: Journey to the center of the knowledge neurons: discoveries of language-independent knowledge neurons and degenerate knowledge neurons. In: Proceedings of the 38th AAAI Conference on Artificial Intelligence, vol.\u00a038, pp. 17817\u201317825 (2024)","DOI":"10.1609\/aaai.v38i16.29735"},{"key":"14_CR7","doi-asserted-by":"crossref","unstructured":"Dai, D., Dong, L., Hao, Y., Sui, Z., Chang, B., Wei, F.: Knowledge neurons in pretrained transformers. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 8493\u20138502 (2022)","DOI":"10.18653\/v1\/2022.acl-long.581"},{"key":"14_CR8","unstructured":"Das, R., et al.: Go for a walk and arrive at the answer: reasoning over paths in knowledge bases using reinforcement learning. In: International Conference on Learning Representations (2018)"},{"key":"14_CR9","doi-asserted-by":"crossref","unstructured":"Dettmers, T., Minervini, P., Stenetorp, P., Riedel, S.: Convolutional 2D knowledge graph embeddings. In: Proceedings of the 32nd AAAI Conference on Artificial Intelligence, vol.\u00a032, pp. 1811\u20131818 (2018)","DOI":"10.1609\/aaai.v32i1.11573"},{"key":"14_CR10","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186 (2019)"},{"key":"14_CR11","unstructured":"Dong, Y., Cordonnier, J.B., Loukas, A.: Attention is not all you need: pure attention loses rank doubly exponentially with depth. In: Proceedings of the 38th International Conference on Machine Learning, pp. 2793\u20132803. PMLR (2021)"},{"key":"14_CR12","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"14_CR13","doi-asserted-by":"crossref","unstructured":"Geva, M., Schuster, R., Berant, J., Levy, O.: Transformer feed-forward layers are key-value memories. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 5484\u20135495 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.446"},{"key":"14_CR14","doi-asserted-by":"crossref","unstructured":"Hao, Y., Dong, L., Wei, F., Xu, K.: Self-attention attribution: interpreting information interactions inside transformer. In: Proceedings of the 35th AAAI Conference on Artificial Intelligence, vol.\u00a035, pp. 12963\u201312971 (2021)","DOI":"10.1609\/aaai.v35i14.17533"},{"issue":"4","key":"14_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3447772","volume":"54","author":"A Hogan","year":"2021","unstructured":"Hogan, A., et al.: Knowledge graphs. ACM Comput. Surv. 54(4), 1\u201337 (2021)","journal-title":"ACM Comput. Surv."},{"key":"14_CR16","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1162\/tacl_a_00324","volume":"8","author":"Z Jiang","year":"2020","unstructured":"Jiang, Z., Xu, F.F., Araki, J., Neubig, G.: How can we know what language models know? Trans. Assoc. Comput. Linguist. 8, 423\u2013438 (2020)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"14_CR17","unstructured":"Jiang, Z., et al.: Gradient rewiring for editable graph neural network training. In: Proceedings of the 38th Annual Conference on Neural Information Processing Systems (2024)"},{"key":"14_CR18","unstructured":"Kipf, T.N., Welling, M.: Variational graph auto-encoders. arXiv preprint arXiv:1611.07308 (2016)"},{"key":"14_CR19","doi-asserted-by":"publisher","first-page":"188","DOI":"10.1007\/978-981-99-1645-0_16","volume-title":"Neural Information Processing: 29th International Conference, ICONIP 2022, Virtual Event, November 22\u201326, 2022, Proceedings, Part VI","author":"J Li","year":"2023","unstructured":"Li, J., Zhang, K., Pu, X., Kong, Y.: Graph attention mixup transformer for\u00a0graph classification. In: Tanveer, M., Agarwal, S., Ozawa, S., Ekbal, A., Jatowt, A. (eds.) Neural Information Processing: 29th International Conference, ICONIP 2022, Virtual Event, November 22\u201326, 2022, Proceedings, Part VI, pp. 188\u2013199. Springer Nature Singapore, Singapore (2023). https:\/\/doi.org\/10.1007\/978-981-99-1645-0_16"},{"key":"14_CR20","doi-asserted-by":"crossref","unstructured":"Liu, C., et al.: Gapformer: graph transformer with graph pooling for node classification. In: Proceedings of the 32nd International Joint Conferences on Artificial Intelligence Organization, pp. 2196\u20132205 (2023)","DOI":"10.24963\/ijcai.2023\/244"},{"key":"14_CR21","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the 2021 IEEE\/CVF International Conference on Computer Vision, pp. 9992\u201310002 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"14_CR22","unstructured":"Liu, Z., et al.: Editable graph neural network for node classifications. In: Proceedings of the 19th International Workshop on Mining and Learning with Graphs (2023)"},{"key":"14_CR23","unstructured":"Mika, G.P.: Toward a transparent recommender system. In: Proceedings of the 14th International Rule Challenge, 4th Doctoral Consortium, and 6th Industry Track @ RuleML+RR, vol.\u00a02644, pp. 111\u2013119 (2020)"},{"key":"14_CR24","doi-asserted-by":"crossref","unstructured":"Mika, G.P., Bouzeghoub, A., W\u0119grzyn-Wolska, K., Neggaz, Y.M.: HGExplainer: explainable graph neural network. In: Proceedings of the 22nd IEEE\/WIC International Conference on Web Intelligence and Intelligent Agent Technology, pp. 221\u2013229 (2023)","DOI":"10.1109\/WI-IAT59888.2023.00035"},{"issue":"1","key":"14_CR25","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1109\/JPROC.2015.2483592","volume":"104","author":"M Nickel","year":"2015","unstructured":"Nickel, M., Murphy, K., Tresp, V., Gabrilovich, E.: A review of relational machine learning for knowledge graphs. Proc. IEEE 104(1), 11\u201333 (2015)","journal-title":"Proc. IEEE"},{"key":"14_CR26","doi-asserted-by":"crossref","unstructured":"Pahuja, V., Wang, B., Latapie, H., Srinivasa, J., Su, Y.: A retrieve-and-read framework for knowledge graph link prediction. In: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, pp. 1992\u20132002 (2023)","DOI":"10.1145\/3583780.3614769"},{"key":"14_CR27","unstructured":"Pan, L., Shi, C., Dokmani\u0107, I.: Neural link prediction with walk pooling. In: International Conference on Learning Representations (2022)"},{"key":"14_CR28","doi-asserted-by":"crossref","unstructured":"Park, D.H., et al.: Multimodal explanations: justifying decisions and pointing to the evidence. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8779\u20138788 (2018)","DOI":"10.1109\/CVPR.2018.00915"},{"key":"14_CR29","doi-asserted-by":"crossref","unstructured":"Petroni, F., et al.: Language models as knowledge bases? In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pp. 2463\u20132473 (2019)","DOI":"10.18653\/v1\/D19-1250"},{"key":"14_CR30","doi-asserted-by":"crossref","unstructured":"Ross, A.S., Hughes, M.C., Doshi-Velez, F.: Right for the right reasons: training differentiable models by constraining their explanations. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence, pp. 2662\u20132670 (2017)","DOI":"10.24963\/ijcai.2017\/371"},{"key":"14_CR31","doi-asserted-by":"crossref","unstructured":"Shomer, H., Ma, Y., Mao, H., Li, J., Wu, B., Tang, J.: LPFormer: an adaptive graph transformer for link prediction. In: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 2686\u20132698 (2024)","DOI":"10.1145\/3637528.3672025"},{"key":"14_CR32","unstructured":"Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: Proceedings of the 34th International Conference on Machine Learning, pp. 3319\u20133328 (2017)"},{"key":"14_CR33","doi-asserted-by":"crossref","unstructured":"Tenney, I., Das, D., Pavlick, E.: BERT rediscovers the classical NLP pipeline. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 4593\u20134601 (2019)","DOI":"10.18653\/v1\/P19-1452"},{"key":"14_CR34","doi-asserted-by":"crossref","unstructured":"Toutanova, K., Chen, D.: Observed versus latent features for knowledge base and text inference. In: Proceedings of the 3rd Workshop on Continuous Vector Space Models and Their Compositionality, pp. 57\u201366 (2015)","DOI":"10.18653\/v1\/W15-4007"},{"key":"14_CR35","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"issue":"8","key":"14_CR36","doi-asserted-by":"publisher","first-page":"961","DOI":"10.3390\/sym16080961","volume":"16","author":"D Wang","year":"2024","unstructured":"Wang, D., et al.: MM-transformer: a transformer-based knowledge graph link prediction model that fuses multimodal features. Symmetry 16(8), 961 (2024)","journal-title":"Symmetry"},{"issue":"12","key":"14_CR37","doi-asserted-by":"publisher","first-page":"2724","DOI":"10.1109\/TKDE.2017.2754499","volume":"29","author":"Q Wang","year":"2017","unstructured":"Wang, Q., Mao, Z., Wang, B., Guo, L.: Knowledge graph embedding: a survey of approaches and applications. IEEE Trans. Knowl. Data Eng. 29(12), 2724\u20132743 (2017)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"14_CR38","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Learning intents behind interactions with knowledge graph for recommendation. In: Proceedings of the 30th Web Conference 2021, pp. 878\u2013887 (2021)","DOI":"10.1145\/3442381.3450133"},{"key":"14_CR39","unstructured":"Wu, F., Fan, A., Baevski, A., Dauphin, Y.N., Auli, M.: Pay less attention with lightweight and dynamic convolutions. In: International Conference on Learning Representations (2019)"},{"key":"14_CR40","doi-asserted-by":"publisher","first-page":"106194","DOI":"10.1016\/j.knosys.2020.106194","volume":"204","author":"Z Yang","year":"2020","unstructured":"Yang, Z., Dong, S.: HAGERec: hierarchical attention graph convolutional network incorporating knowledge graph for explainable recommendation. Knowl. Based Syst. 204, 106194 (2020)","journal-title":"Knowl. Based Syst."},{"key":"14_CR41","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1016\/j.patrec.2024.02.014","volume":"180","author":"Z Zhang","year":"2024","unstructured":"Zhang, Z., et al.: GBCA: graph convolution network and BERT combined with co-attention for fake news detection. Pattern Recogn. Lett. 180, 26\u201332 (2024)","journal-title":"Pattern Recogn. Lett."},{"key":"14_CR42","unstructured":"Zhong, S., et al.: GNNs also deserve editing, and they need it more than once. In: Proceedings of the 41st International Conference on Machine Learning, vol.\u00a0235, pp. 61727\u201361746 (2024)"}],"container-title":["Lecture Notes in Computer Science","SOFSEM 2025: Theory and Practice of Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-82697-9_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,15]],"date-time":"2025-02-15T10:17:48Z","timestamp":1739614668000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-82697-9_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031826962","9783031826979"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-82697-9_14","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"16 February 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"SOFSEM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Current Trends in Theory and Practice of Computer Science","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bratislava","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Slovakia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 January 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 January 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"50","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"sofsem2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.sofsem.sk","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}