{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T13:30:45Z","timestamp":1767706245810,"version":"3.48.0"},"publisher-location":"Singapore","reference-count":48,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819541577","type":"print"},{"value":"9789819541584","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-4158-4_14","type":"book-chapter","created":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T02:31:30Z","timestamp":1767321090000},"page":"220-235","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Explaining Black-Box Language Models with\u00a0Knowledge Probing Systems: A Post-hoc Explanation Perspective"],"prefix":"10.1007","author":[{"given":"Yunxiao","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Hao","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Zhiqiang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xiaoli","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jiye","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Ru","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,2]]},"reference":[{"key":"14_CR1","doi-asserted-by":"crossref","unstructured":"Baker, C.F., Fillmore, C.J., Lowe, J.B.: The Berkeley FrameNet project. In: Proceedings of the ACL (1998)","DOI":"10.3115\/980451.980860"},{"key":"14_CR2","doi-asserted-by":"crossref","unstructured":"Bouraoui, Z., Camacho-Collados, J., Schockaert, S.: Inducing relational knowledge from BERT. In: Proceedings of the AAAI (2020)","DOI":"10.1609\/aaai.v34i05.6242"},{"key":"14_CR3","doi-asserted-by":"crossref","unstructured":"Chen, Y., Zhong, R., Zha, S., Karypis, G., He, H.: Meta-learning via language model in-context tuning. In: Proceedings of the ACL (2022)","DOI":"10.18653\/v1\/2022.acl-long.53"},{"key":"14_CR4","doi-asserted-by":"crossref","unstructured":"Cui, Y., Che, W., Liu, T., Qin, B., Wang, S., Hu, G.: Revisiting pre-trained models for Chinese natural language processing. In: Proceedings of the EMNLP (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.58"},{"key":"14_CR5","doi-asserted-by":"crossref","unstructured":"De\u00a0Cao, N., Aziz, W., Titov, I.: Editing factual knowledge in language models. In: Proceedings of the EMNLP (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.522"},{"key":"14_CR6","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the Conference of the NAACL (2019)"},{"key":"14_CR7","doi-asserted-by":"crossref","unstructured":"Du, Z., et al.: GLM: general language model pretraining with autoregressive blank infilling. In: Proceedings of the ACL (2022)","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"14_CR8","unstructured":"Dubey, A., et\u00a0al.: The Llama 3 Herd of models. arXiv preprint arXiv:2407.21783 (2024)"},{"key":"14_CR9","doi-asserted-by":"crossref","unstructured":"Dufter, P., Kassner, N., Sch\u00fctze, H.: Static embeddings as efficient knowledge bases? In: Proceedings of the NAACL (2021)","DOI":"10.18653\/v1\/2021.naacl-main.186"},{"key":"14_CR10","doi-asserted-by":"crossref","unstructured":"Haviv, A., Berant, J., Globerson, A.: BERTese: learning to speak to BERT. In: Proceedings of the Conference of the EACL (2021)","DOI":"10.18653\/v1\/2021.eacl-main.316"},{"key":"14_CR11","unstructured":"Jiang, A.Q., et\u00a0al.: Mixtral of experts. arXiv preprint arXiv:2401.04088 (2024)"},{"key":"14_CR12","doi-asserted-by":"crossref","unstructured":"Jiang, Z., Xu, F.F., Araki, J., Neubig, G.: How can we know what language models know? Trans. Assoc. Comput. Linguist. (2020)","DOI":"10.1162\/tacl_a_00324"},{"key":"14_CR13","unstructured":"Kalo, J.C.: KAMEL: knowledge analysis with multitoken entities in language models. In: Conference on Automated Knowledge Base Construction (2022)"},{"key":"14_CR14","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Bengio, Y., LeCun, Y. (eds.) Proceedings of the ICLR (2015)"},{"key":"14_CR15","doi-asserted-by":"crossref","unstructured":"Lee, K., Han, W., Hwang, S.w., Lee, H., Park, J., Lee, S.W.: Plug-and-play adaptation for continuously-updated QA. In: Findings of the ACL 2022 (2022)","DOI":"10.18653\/v1\/2022.findings-acl.37"},{"key":"14_CR16","doi-asserted-by":"crossref","unstructured":"Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., et\u00a0al: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the ACL (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"14_CR17","doi-asserted-by":"crossref","unstructured":"Lewis, P., Stenetorp, P., Riedel, S.: Question and answer test-train overlap in open-domain question answering datasets. In: Proceedings of the Conference of the EACL (2021)","DOI":"10.18653\/v1\/2021.eacl-main.86"},{"issue":"2","key":"14_CR18","doi-asserted-by":"publisher","first-page":"239","DOI":"10.1007\/s11633-023-1450-8","volume":"21","author":"R Li","year":"2024","unstructured":"Li, R., et al.: A comprehensive overview of CFN from a commonsense perspective. Mach. Intell. Res. 21(2), 239\u2013256 (2024)","journal-title":"Mach. Intell. Res."},{"key":"14_CR19","doi-asserted-by":"crossref","unstructured":"Li\u00e9tard, B., Abdou, M., S\u00f8gaard, A.: Do language models know the way to Rome? In: Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (2021)","DOI":"10.18653\/v1\/2021.blackboxnlp-1.40"},{"key":"14_CR20","doi-asserted-by":"crossref","unstructured":"Mallen, A., Asai, A., Zhong, V., Das, R., Khashabi, D., Hajishirzi, H.: When not to trust language models: investigating effectiveness of parametric and non-parametric memories. In: Proceedings of the ACL (2023)","DOI":"10.18653\/v1\/2023.acl-long.546"},{"key":"14_CR21","doi-asserted-by":"crossref","unstructured":"Margatina, K., Wang, S., Vyas, Y., Anna\u00a0John, N., Benajiba, Y., Ballesteros, M.: Dynamic benchmarking of masked language models on temporal concept drift with multiple views. In: Proceedings of the Conference of the EACL (2023)","DOI":"10.18653\/v1\/2023.eacl-main.211"},{"key":"14_CR22","unstructured":"Meng, K., Bau, D., Andonian, A., Belinkov, Y.: Locating and editing factual associations in GPT. Adv. Neural Inf. Process. Syst. (2022)"},{"key":"14_CR23","doi-asserted-by":"crossref","unstructured":"Meng, Z., Liu, F., Shareghi, E., Su, Y., Collins, C., Collier, N.: Rewire-then-probe: a contrastive recipe for probing biomedical knowledge of pre-trained language models. In: Proceedings of the ACL (2022)","DOI":"10.18653\/v1\/2022.acl-long.329"},{"key":"14_CR24","unstructured":"OpenAI: Chat-GPT (2023). Accessed Oct 2023"},{"key":"14_CR25","doi-asserted-by":"crossref","unstructured":"Petroni, F., et al.: Language models as knowledge bases? In: Proceedings of the EMNLP (2019)","DOI":"10.18653\/v1\/D19-1250"},{"key":"14_CR26","doi-asserted-by":"crossref","unstructured":"Pet\u00f6fi, J.S.: A frame for frames. In: Proceedings of the 2nd Annual Meeting of the Berkeley Linguistics Society (1976)","DOI":"10.3765\/bls.v2i0.2271"},{"key":"14_CR27","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100,000+ questions for machine comprehension of text. In: Proceedings of the EMNLP (2016)","DOI":"10.18653\/v1\/D16-1264"},{"key":"14_CR28","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: \u201cwhy should i trust you?\u201d: Explaining the predictions of any classifier. In: Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (SIGKDD) (2016)","DOI":"10.1145\/2939672.2939778"},{"key":"14_CR29","doi-asserted-by":"crossref","unstructured":"Roberts, A., Raffel, C., Shazeer, N.: How much knowledge can you pack into the parameters of a language model? In: Proceedings of the EMNLP (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.437"},{"key":"14_CR30","doi-asserted-by":"crossref","unstructured":"Saeed, M., Papotti, P.: You are my type! Type embeddings for pre-trained language models. In: Findings of the EMNLP 2022 (2022)","DOI":"10.18653\/v1\/2022.findings-emnlp.336"},{"key":"14_CR31","doi-asserted-by":"crossref","unstructured":"Shi, W., Joshi, M., Zettlemoyer, L.: DESCGEN: a distantly supervised dataset for generating entity descriptions. In: Proceedings of the ACL (2021)","DOI":"10.18653\/v1\/2021.acl-long.35"},{"key":"14_CR32","unstructured":"Su, X., et al.: A knowledge-guided framework for frame identification. In: Proceedings of the ACL (2021)"},{"key":"14_CR33","doi-asserted-by":"crossref","unstructured":"Sun, K., Yu, D., Yu, D., Cardie, C.: Investigating prior knowledge for challenging Chinese machine reading comprehension. Trans. Assoc. Comput. Linguist. (TACL) (2020)","DOI":"10.1162\/tacl_a_00305"},{"key":"14_CR34","unstructured":"Sun, Y., et al.: ERNIE 3.0: large-scale knowledge enhanced pre-training for language understanding and generation. CoRR (2021)"},{"key":"14_CR35","doi-asserted-by":"crossref","unstructured":"Sung, M., Lee, J., Yi, S., Jeon, M., Kim, S., Kang, J.: Can language models be biomedical knowledge bases? In: Proceedings of the EMNLP (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.388"},{"key":"14_CR36","doi-asserted-by":"crossref","unstructured":"Tan, H., et al.: GCRC: a new challenging MRC dataset from Gaokao Chinese for explainable evaluation. In: Findings of the ACL 2021 (2021)","DOI":"10.18653\/v1\/2021.findings-acl.113"},{"key":"14_CR37","unstructured":"Team, G., et\u00a0al.: Gemma: open models based on Gemini research and technology. arXiv preprint arXiv:2403.08295 (2024)"},{"key":"14_CR38","unstructured":"Touvron, H., et\u00a0al.: Llama 2: open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)"},{"key":"14_CR39","doi-asserted-by":"crossref","unstructured":"Wang, C., Liu, P., Zhang, Y.: Can generative pre-trained language models serve as knowledge bases for closed-book QA? In: Proceedings of the ACL (2021)","DOI":"10.18653\/v1\/2021.acl-long.251"},{"key":"14_CR40","doi-asserted-by":"crossref","unstructured":"Xue, N., Xia, F., Chiou, F.D., Palmer, M.: The Penn Chinese treebank: phrase structure annotation of a large corpus. Nat. Lang. Eng. (2005)","DOI":"10.1017\/S135132490400364X"},{"key":"14_CR41","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems (NeurIPS) (2019)"},{"key":"14_CR42","doi-asserted-by":"crossref","unstructured":"Ye, J., et al.: ZeroGen: efficient zero-shot learning via dataset generation. In: Proceedings of the EMNLP (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.801"},{"key":"14_CR43","doi-asserted-by":"crossref","unstructured":"Youssef, P., Kora\u015f, O., Li, M., Schl\u00f6tterer, J., Seifert, C.: Give me the facts! A survey on factual knowledge probing in pre-trained language models. In: Findings of the EMNLP 2023 (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.1043"},{"key":"14_CR44","doi-asserted-by":"crossref","unstructured":"Yuan, H., Cai, L., Hu, X., et\u00a0al.: Interpreting image classifiers by generating discrete masks. IEEE Trans. Pattern Anal. Mach. Intell. (2020)","DOI":"10.1109\/TPAMI.2020.3028783"},{"key":"14_CR45","doi-asserted-by":"crossref","unstructured":"Zhang, X., Yang, Q.: Xuanyuan 2.0: a large Chinese financial chat model with hundreds of billions parameters. In: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (2023)","DOI":"10.1145\/3583780.3615285"},{"key":"14_CR46","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Fei, H., Li, D., Li, P.: PromptGen: automatically generate prompts using generative models. In: Findings of the NAACL 2022 (2022)","DOI":"10.18653\/v1\/2022.findings-naacl.3"},{"key":"14_CR47","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Wang, Z., Li, X., Liang, J., Li, R.: AGR: reinforced causal agent-guided self-explaining rationalization. In: Proceedings of the ACL (2024)","DOI":"10.18653\/v1\/2024.acl-short.47"},{"key":"14_CR48","doi-asserted-by":"crossref","unstructured":"Zheng, C., Wang, Y., Chang, B.: Query your model with definitions in framenet: an effective method for frame semantic role labeling. In: Proceedings of the AAAI (2023)","DOI":"10.1609\/aaai.v37i11.26642"}],"container-title":["Lecture Notes in Computer Science","Database Systems for Advanced Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-4158-4_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T02:31:33Z","timestamp":1767321093000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-4158-4_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819541577","9789819541584"],"references-count":48,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-4158-4_14","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"2 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DASFAA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Database Systems for Advanced Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 May 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 May 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dasfaa2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/dasfaa2025.github.io","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}