{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T18:41:27Z","timestamp":1762368087902,"version":"build-2065373602"},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,8,31]],"date-time":"2025-08-31T00:00:00Z","timestamp":1756598400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,8,31]],"date-time":"2025-08-31T00:00:00Z","timestamp":1756598400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,8,31]]},"DOI":"10.1109\/mlsp62443.2025.11204312","type":"proceedings-article","created":{"date-parts":[[2025,10,24]],"date-time":"2025-10-24T17:15:52Z","timestamp":1761326152000},"page":"1-6","source":"Crossref","is-referenced-by-count":0,"title":["Tackling Distribution Shift in LLM via KILO: Knowledge-Instructed Learning for Continual Adaptation"],"prefix":"10.1109","author":[{"given":"Ling","family":"Muttakhiroh","sequence":"first","affiliation":[{"name":"Concordia University,Montreal,Canada"}]},{"given":"Thomas","family":"Fevens","sequence":"additional","affiliation":[{"name":"Concordia University,Montreal,Canada"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Error sensitivity modulation based experience replay: Mitigating abrupt representation drift in continual learning","author":"Sarfraz","year":"2023","journal-title":"arXiv preprint"},{"key":"ref2","article-title":"Continual pretraining of large language models: How to (re) warm your model?","author":"Gupta","year":"2023","journal-title":"arXiv preprint"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2024.106492"},{"key":"ref4","article-title":"Recyclable tuning for continual pretraining","author":"Qin","year":"2023","journal-title":"arXiv preprint"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3735633"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN52387.2021.9533862"},{"key":"ref7","article-title":"Graph relation aware continual learning","author":"Shen","year":"2023","journal-title":"arXiv preprint"},{"key":"ref8","article-title":"Continual graph learning: A survey","author":"Yuan","year":"2023","journal-title":"arXiv preprint"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.33"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.244"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00681"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-024-07930-y"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.85"},{"key":"ref14","first-page":"13006","article-title":"CGLB: Benchmark tasks for continual graph learning","volume":"35","author":"Zhang","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref15","article-title":"Continual learning on graphs: Challenges, solutions, and opportunities","author":"Zhang","year":"2024","journal-title":"arXiv preprint"},{"key":"ref16","article-title":"Knowledge enhanced prompt tuning for few-shot learning","author":"Gu","year":"2022","journal-title":"Findings of ACL"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_37"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6428"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.351"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.695"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i5.16602"},{"key":"ref22","article-title":"Unifying knowledge graph learning and reasoning: A roadmap","author":"Pan","year":"2020","journal-title":"arXiv preprint"},{"key":"ref23","article-title":"Graph2text or Graph2token: A perspective of large language models for graph learning","author":"Yu","year":"2025","journal-title":"arXiv preprint"},{"key":"ref24","article-title":"Learning on graphs with large language models (LLMs): A deep dive into model robustness","author":"Guo","year":"2024","journal-title":"arXiv preprint"},{"key":"ref25","article-title":"Pointer sentinel mixture models","author":"Merity","year":"2016","journal-title":"arXiv preprint"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1038\/s41597-023-02068-4"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W17-4413"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.148"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.331"},{"issue":"140","key":"ref30","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of machine learning research"}],"event":{"name":"2025 IEEE 35th International Workshop on Machine Learning for Signal Processing (MLSP)","start":{"date-parts":[[2025,8,31]]},"location":"Istanbul, Turkiye","end":{"date-parts":[[2025,9,3]]}},"container-title":["2025 IEEE 35th International Workshop on Machine Learning for Signal Processing (MLSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11204201\/11204202\/11204312.pdf?arnumber=11204312","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T18:37:13Z","timestamp":1762367833000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11204312\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,31]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/mlsp62443.2025.11204312","relation":{},"subject":[],"published":{"date-parts":[[2025,8,31]]}}}