{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T01:54:54Z","timestamp":1743040494952,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":31,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819611478"},{"type":"electronic","value":"9789819611485"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-1148-5_11","type":"book-chapter","created":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T13:44:40Z","timestamp":1737121480000},"page":"131-142","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Collaborative Framework for\u00a0Dynamic Knowledge Updating and\u00a0Transparent Reasoning with\u00a0Large Language Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-2139-8674","authenticated-orcid":false,"given":"Ziyu","family":"Ding","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9570-7566","authenticated-orcid":false,"given":"Pei-Gen","family":"Ye","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0003-4750-0192","authenticated-orcid":false,"given":"Yaqi","family":"Wu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-5823-8343","authenticated-orcid":false,"given":"Huali","family":"Ren","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,18]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Baek, J., Aji, A.F., Saffari, A.: Knowledge-augmented language model prompting for zero-shot knowledge graph question answering. arXiv preprint arXiv:2306.04136 (2023)","key":"11_CR1","DOI":"10.18653\/v1\/2023.nlrse-1.7"},{"doi-asserted-by":"crossref","unstructured":"Bang, Y., et\u00a0al.: A multitask, multilingual, multimodal evaluation of ChatGPT on reasoning, hallucination, and interactivity. arXiv preprint arXiv:2302.04023 (2023)","key":"11_CR2","DOI":"10.18653\/v1\/2023.ijcnlp-main.45"},{"doi-asserted-by":"crossref","unstructured":"Betz, P., Meilicke, C., Stuckenschmidt, H.: Adversarial explanations for knowledge graph embeddings. In: IJCAI, vol.\u00a02022, pp. 2820\u20132826 (2022)","key":"11_CR3","DOI":"10.24963\/ijcai.2022\/391"},{"unstructured":"Brown, T.B.: Language models are few-shot learners. arXiv preprint arXiv:2005.14165 (2020)","key":"11_CR4"},{"issue":"3","key":"11_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3641289","volume":"15","author":"Y Chang","year":"2024","unstructured":"Chang, Y., et al.: A survey on evaluation of large language models. ACM Trans. Intell. Syst. Technol. 15(3), 1\u201345 (2024)","journal-title":"ACM Trans. Intell. Syst. Technol."},{"unstructured":"Devlin, J.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)","key":"11_CR6"},{"unstructured":"Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. In: Advances in Neural Information Processing Systems, vol. 30 (2017)","key":"11_CR7"},{"doi-asserted-by":"crossref","unstructured":"He, G., Lan, Y., Jiang, J., Zhao, W.X., Wen, J.R.: Improving multi-hop knowledge base question answering by learning intermediate supervision signals. In: Proceedings of the 14th ACM International Conference on Web Search and Data Mining, pp. 553\u2013561 (2021)","key":"11_CR8","DOI":"10.1145\/3437963.3441753"},{"unstructured":"He, R., et al.: On the effectiveness of adapter-based tuning for pretrained language model adaptation. arXiv preprint arXiv:2106.03164 (2021)","key":"11_CR9"},{"unstructured":"Hu, E.J., et al.: LoRA: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)","key":"11_CR10"},{"unstructured":"Jiang, J., Zhou, K., Zhao, W.X., Wen, J.R.: UniKGQA: unified retrieval and reasoning for solving multi-hop question answering over knowledge graph. arXiv preprint arXiv:2212.00959 (2022)","key":"11_CR11"},{"unstructured":"Kaddour, J., Harris, J., Mozes, M., Bradley, H., Raileanu, R., McHardy, R.: Challenges and applications of large language models. arXiv preprint arXiv:2307.10169 (2023)","key":"11_CR12"},{"doi-asserted-by":"crossref","unstructured":"Lester, B., Al-Rfou, R., Constant, N.: The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691 (2021)","key":"11_CR13","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"doi-asserted-by":"crossref","unstructured":"Li, F., Chen, M., Dong, R.: Multi-hop question answering with knowledge graph embedding in a similar semantic space. In: 2022 International Joint Conference on Neural Networks (IJCNN), pp. 01\u201307. IEEE (2022)","key":"11_CR14","DOI":"10.1109\/IJCNN55064.2022.9892550"},{"doi-asserted-by":"crossref","unstructured":"Li, S., et al.: Graph reasoning for question answering with triplet retrieval. arXiv preprint arXiv:2305.18742 (2023)","key":"11_CR15","DOI":"10.18653\/v1\/2023.findings-acl.208"},{"doi-asserted-by":"crossref","unstructured":"Li, T., Ma, X., Zhuang, A., Gu, Y., Su, Y., Chen, W.: Few-shot in-context learning for knowledge base question answering. arXiv preprint arXiv:2305.01750 (2023)","key":"11_CR16","DOI":"10.18653\/v1\/2023.acl-long.385"},{"doi-asserted-by":"crossref","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190 (2021)","key":"11_CR17","DOI":"10.18653\/v1\/2021.acl-long.353"},{"doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: P-Tuning v2: prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602 (2021)","key":"11_CR18","DOI":"10.18653\/v1\/2022.acl-short.8"},{"unstructured":"Luo, L., Li, Y.F., Haffari, G., Pan, S.: Reasoning on graphs: faithful and interpretable large language model reasoning. arXiv preprint arXiv:2310.01061 (2023)","key":"11_CR19"},{"doi-asserted-by":"crossref","unstructured":"Pan, S., Luo, L., Wang, Y., Chen, C., Wang, J., Wu, X.: Unifying large language models and knowledge graphs: a roadmap. IEEE Trans. Knowl. Data Eng. (2024)","key":"11_CR20","DOI":"10.1109\/TKDE.2024.3352100"},{"doi-asserted-by":"crossref","unstructured":"Saxena, A., Tripathi, A., Talukdar, P.: Improving multi-hop question answering over knowledge graphs using knowledge base embeddings. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 4498\u20134507 (2020)","key":"11_CR21","DOI":"10.18653\/v1\/2020.acl-main.412"},{"doi-asserted-by":"crossref","unstructured":"Sun, H., Bedrax-Weiss, T., Cohen, W.W.: PullNet: open domain question answering with iterative retrieval on knowledge bases and text. arXiv preprint arXiv:1904.09537 (2019)","key":"11_CR22","DOI":"10.18653\/v1\/D19-1242"},{"doi-asserted-by":"crossref","unstructured":"Sun, Y., Zhang, L., Cheng, G., Qu, Y.: SPARQA: skeleton-based semantic parsing for complex questions over knowledge bases. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a034, pp. 8952\u20138959 (2020)","key":"11_CR23","DOI":"10.1609\/aaai.v34i05.6426"},{"issue":"4","key":"11_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3543856","volume":"18","author":"Z Sun","year":"2022","unstructured":"Sun, Z., Yang, H., Liu, K., Yin, Z., Li, Z., Xu, W.: Recent advances in LoRA: a comprehensive survey. ACM Trans. Sens. Netw. 18(4), 1\u201344 (2022)","journal-title":"ACM Trans. Sens. Netw."},{"unstructured":"Touvron, H., et\u00a0al.: Llama 2: open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)","key":"11_CR25"},{"unstructured":"Vos, D., D\u00f6hmen, T., Schelter, S.: Towards parameter-efficient automation of data wrangling tasks with prefix-tuning. In: NeurIPS 2022 First Table Representation Workshop (2022)","key":"11_CR26"},{"unstructured":"Xu, Z., Jain, S., Kankanhalli, M.: Hallucination is inevitable: An innate limitation of large language models. arXiv preprint arXiv:2401.11817 (2024)","key":"11_CR27"},{"unstructured":"Yu, D., et al.: DecAF: joint decoding of answers and logical forms for question answering over knowledge bases. arXiv preprint arXiv:2210.00063 (2022)","key":"11_CR28"},{"doi-asserted-by":"crossref","unstructured":"Zhang, J., et al.: Subgraph retrieval enhanced model for multi-hop knowledge base question answering. arXiv preprint arXiv:2202.13296 (2022)","key":"11_CR29","DOI":"10.18653\/v1\/2022.acl-long.396"},{"unstructured":"Zhang, R., et al.: Llama-adapter: efficient fine-tuning of language models with zero-init attention. arXiv preprint arXiv:2303.16199 (2023)","key":"11_CR30"},{"unstructured":"Zhou, Y., et al.: Large language models are human-level prompt engineers. arXiv preprint arXiv:2211.01910 (2022)","key":"11_CR31"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence Security and Privacy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-1148-5_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T13:44:51Z","timestamp":1737121491000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-1148-5_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819611478","9789819611485"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-1148-5_11","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"18 January 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"AIS&P","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Intelligence Security and Privacy","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Guangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ais&p2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}