{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T04:27:01Z","timestamp":1750134421703,"version":"3.28.0"},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,11,18]],"date-time":"2023-11-18T00:00:00Z","timestamp":1700265600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,11,18]],"date-time":"2023-11-18T00:00:00Z","timestamp":1700265600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,11,18]]},"DOI":"10.1109\/ialp61005.2023.10337079","type":"proceedings-article","created":{"date-parts":[[2023,12,12]],"date-time":"2023-12-12T18:34:31Z","timestamp":1702406071000},"page":"258-264","source":"Crossref","is-referenced-by-count":2,"title":["Long-Term Memory for Large Language Models Through Topic-Based Vector Database"],"prefix":"10.1109","author":[{"given":"Yi","family":"Zhang","sequence":"first","affiliation":[{"name":"Koiverse.AI,Hangzhou,China"}]},{"given":"Zhongyang","family":"Yu","sequence":"additional","affiliation":[{"name":"Koiverse.AI,Hangzhou,China"}]},{"given":"Wanqi","family":"Jiang","sequence":"additional","affiliation":[{"name":"Koiverse.AI,Hangzhou,China"}]},{"given":"Yufeng","family":"Shen","sequence":"additional","affiliation":[{"name":"Koiverse.AI,Hangzhou,China"}]},{"given":"Jin","family":"Li","sequence":"additional","affiliation":[{"name":"Koiverse.AI,Hangzhou,China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"27 730","article-title":"Training language models to follow instructions with human feedback","volume":"35","author":"Ouyang","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref2","first-page":"1877","article-title":"Language mod-els are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref3","article-title":"Multitask prompted training enables zero-shot task generalization","author":"Sanh","year":"2021","journal-title":"ar Xiv preprint"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.759"},{"article-title":"Instructgpt: Ai for generating instructions","volume-title":"OpenAI","year":"2023","key":"ref5"},{"key":"ref6","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"ar Xiv preprint"},{"key":"ref7","article-title":"Palm: Scaling language modeling with pathways","author":"Chowdhery","year":"2022","journal-title":"arXiv preprint"},{"key":"ref8","article-title":"A survey of large language models","author":"Zhao","year":"2023","journal-title":"arXiv preprint"},{"key":"ref9","article-title":"Sparks of artificial general intelligence: Early experiments with gpt-4","author":"Bubeck","year":"2023","journal-title":"arXiv preprint"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.lindif.2023.102274"},{"key":"ref11","article-title":"Scaling instruction-finetuned language models","author":"Chung","year":"2022","journal-title":"arXiv preprint"},{"issue":"1","key":"ref12","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"The Journal of Machine Learning Research"},{"key":"ref13","article-title":"Chatlaw: Open-source legal large language model with integrated external knowledge bases","author":"Cui","year":"2023","journal-title":"arXiv preprint"},{"key":"ref14","article-title":"Check your facts and try again: Improving large language models with external knowledge and automated feedback","author":"Peng","year":"2023","journal-title":"arXiv preprint"},{"key":"ref15","article-title":"Structural embeddings of tools for large language models","author":"Unlu","year":"2023","journal-title":"arXiv preprint"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-1212"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-2033"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5446"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/K16-2014"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1004"},{"key":"ref21","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"neural information processing systems"},{"key":"ref22","article-title":"Deep reinforcement learning from human preferences","volume":"30","author":"Christiano","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.iheduc.2021.100817"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-35320-8_1"},{"key":"ref25","article-title":"A prompt pattern catalog to enhance prompt engineering with chatgpt","author":"White","year":"2023","journal-title":"ar Xiv preprint"},{"key":"ref26","article-title":"Tree of thoughts: Deliberate problem solving with large language models","author":"Yao","year":"2023","journal-title":"arXiv preprint"},{"key":"ref27","article-title":"Automatic chain of thought prompting in large language models","author":"Zhang","year":"2022","journal-title":"arXiv preprint"},{"key":"ref28","first-page":"30 583","article-title":"What can transformers learn in-context? a case study of simple function classes","volume":"35","author":"Garg","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref29","article-title":"Self-consistency improves chain of thought reasoning in language models","author":"Wang","year":"2022","journal-title":"arXiv preprint"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.294"},{"key":"ref31","article-title":"Prompting gpt-3 to be reliable","author":"Si","year":"2022","journal-title":"arXiv preprint"},{"key":"ref32","article-title":"word2vec explained: deriving mikolov s negative-sampling word-embedding method","author":"Goldberg","year":"2014","journal-title":"arXiv preprint"},{"journal-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref33"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1410"}],"event":{"name":"2023 International Conference on Asian Language Processing (IALP)","start":{"date-parts":[[2023,11,18]]},"location":"Singapore, Singapore","end":{"date-parts":[[2023,11,20]]}},"container-title":["2023 International Conference on Asian Language Processing (IALP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10336886\/10336974\/10337079.pdf?arnumber=10337079","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,12]],"date-time":"2024-01-12T02:35:28Z","timestamp":1705026928000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10337079\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,18]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/ialp61005.2023.10337079","relation":{},"subject":[],"published":{"date-parts":[[2023,11,18]]}}}