{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,27]],"date-time":"2025-07-27T07:42:59Z","timestamp":1753602179132,"version":"3.28.0"},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,6,30]],"date-time":"2024-06-30T00:00:00Z","timestamp":1719705600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,6,30]],"date-time":"2024-06-30T00:00:00Z","timestamp":1719705600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,6,30]]},"DOI":"10.1109\/ijcnn60899.2024.10650062","type":"proceedings-article","created":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T17:35:05Z","timestamp":1725903305000},"page":"1-8","source":"Crossref","is-referenced-by-count":1,"title":["Beyond Binary Classification: Customizable Text Watermark on Large Language Models"],"prefix":"10.1109","author":[{"given":"Zhenyu","family":"Xu","sequence":"first","affiliation":[{"name":"Texas Tech University,Department of Computer Science,Lubbock,USA"}]},{"given":"Ruoyu","family":"Xu","sequence":"additional","affiliation":[{"name":"Texas Tech University,Department of Computer Science,Lubbock,USA"}]},{"given":"Victor S.","family":"Sheng","sequence":"additional","affiliation":[{"name":"Texas Tech University,Department of Computer Science,Lubbock,USA"}]}],"member":"263","reference":[{"article-title":"The science of detecting LLM-generated texts","year":"2023","author":"Tang","key":"ref1"},{"key":"ref2","article-title":"NYC education department blocks ChatGPT on school devices, networks","volume-title":"Chalkbeat","author":"Elsen-Rooney","year":"2023"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-020-09548-1"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3053998"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/26.24595"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-81897-4_4"},{"article-title":"A watermark for large language models","year":"2023","author":"Kirchenbauer","key":"ref7"},{"key":"ref8","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"J. Mach. Learn. Res."},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"article-title":"OPT: Open pre-trained transformer language models","year":"2022","author":"Zhang","key":"ref10"},{"year":"2024","key":"ref11","article-title":"Babbage-002"},{"year":"2024","key":"ref12","article-title":"Davinci-002"},{"year":"2024","key":"ref13","article-title":"GPT-3.5-Turbo-Instruct"},{"volume-title":"Foundations of Statistical Natural Language Processing.","year":"1999","author":"Manning","key":"ref14"},{"key":"ref15","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"J. Mach. Learn. Res."},{"key":"ref16","article-title":"T5: Text-to-Text Transfer Transformer","volume-title":"Hugging Face","author":"Raffel","year":"2024"},{"article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref17"},{"article-title":"RoBERTa: A robustly optimized BERT pretraining approach","year":"2019","author":"Liu","key":"ref18"},{"key":"ref19","article-title":"Language models are unsupervised multitask learners","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"article-title":"Language models are few-shot learners","volume-title":"Proc. NeurIPS","author":"Brown","key":"ref20"},{"volume-title":"OpenAI Blog","key":"ref21","article-title":"Introducing GPT-4"},{"article-title":"OPT: Open pre-trained transformer language models","year":"2022","author":"Zhang","key":"ref22"},{"volume-title":"Facebook AI Research Blog","key":"ref23","article-title":"LLaMA: Large language models"},{"volume-title":"OpenAI Blog","key":"ref24","article-title":"GPT-3.5: Enhancing language model performance"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-45496-9_14"},{"article-title":"Release strategies and the social impacts of language models","year":"2019","author":"Solaiman","key":"ref26"},{"key":"ref27","article-title":"Detecting texts produced by ChatGPT","author":"Mitrovi\u0107","year":"2023","journal-title":"J. Mach. Learn. Res."},{"key":"ref28","article-title":"GPTZero: Detecting machine-generated texts","author":"Tian","year":"2023","journal-title":"IEEE Trans. Neural Networks and Learning Syst."},{"article-title":"DetectGPT: A tool for identifying GPT-generated texts","year":"2023","author":"Mitchell","key":"ref29"},{"article-title":"On the challenge of detecting GPT-generated texts","volume-title":"Proc. Int. Conf. Mach. Learn","author":"Sadasivan","key":"ref30"}],"event":{"name":"2024 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2024,6,30]]},"location":"Yokohama, Japan","end":{"date-parts":[[2024,7,5]]}},"container-title":["2024 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10649807\/10649898\/10650062.pdf?arnumber=10650062","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,10]],"date-time":"2024-09-10T04:57:53Z","timestamp":1725944273000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10650062\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,30]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/ijcnn60899.2024.10650062","relation":{},"subject":[],"published":{"date-parts":[[2024,6,30]]}}}