{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,19]],"date-time":"2026-01-19T21:32:02Z","timestamp":1768858322512,"version":"3.49.0"},"reference-count":141,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3548967","type":"journal-article","created":{"date-parts":[[2025,3,6]],"date-time":"2025-03-06T18:44:47Z","timestamp":1741286687000},"page":"44707-44721","source":"Crossref","is-referenced-by-count":1,"title":["LLM-Based Text Style Transfer: Have We Taken a Step Forward?"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5045-069X","authenticated-orcid":false,"given":"Martina","family":"Toshevska","sequence":"first","affiliation":[{"name":"Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University in Skopje, Skopje, North Macedonia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3411-2399","authenticated-orcid":false,"given":"Sonja","family":"Gievska","sequence":"additional","affiliation":[{"name":"Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University in Skopje, Skopje, North Macedonia"}]}],"member":"263","reference":[{"key":"ref1","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Sutskever"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.5555\/2969033.2969125"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992696"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1169"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1322"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.inlg-1.25"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11330"},{"key":"ref8","article-title":"Language style transfer from sentences with arbitrary unknown styles","author":"Zhao","year":"2018","journal-title":"arXiv:1808.04071"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/711"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-short.94"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.141"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.inlg-main.27"},{"key":"ref13","article-title":"ProSwitch: Knowledge-guided instruction tuning to switch between professional and non-professional responses","author":"Zong","year":"2024","journal-title":"arXiv:2403.09131"},{"key":"ref14","article-title":"Text style transfer: An introductory overview","author":"Mukherjee","year":"2024","journal-title":"arXiv:2407.14822"},{"key":"ref15","article-title":"A survey of text style transfer: Applications and ethical implications","author":"Mukherjee","year":"2024","journal-title":"arXiv:2407.16737"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1017\/S1351324922000407"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1162\/coli_a_00426"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2021.3115992"},{"key":"ref19","first-page":"1","article-title":"Large language models: A comprehensive survey of its applications, challenges, limitations, and future prospects","volume":"2023","author":"Hadi","year":"2023","journal-title":"Authorea Preprints"},{"key":"ref20","article-title":"A survey of large language models","author":"Xin Zhao","year":"2023","journal-title":"arXiv:2303.18223"},{"key":"ref21","article-title":"Evaluating large language models: A comprehensive survey","author":"Guo","year":"2023","journal-title":"arXiv:2310.19736"},{"key":"ref22","article-title":"Large language models: A survey","author":"Minaee","year":"2024","journal-title":"arXiv:2402.06196"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/3703155"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2024.3469578"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1012"},{"key":"ref26","first-page":"2899","article-title":"Paraphrasing for style","volume-title":"Proc. COLING","author":"Xu"},{"key":"ref27","article-title":"Data-driven approaches for paraphrasing across language variations","author":"Xu","year":"2014"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5385"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.469"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/2623330.2623758"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1145\/2872427.2883037"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.169"},{"key":"ref33","first-page":"311","article-title":"BLEU: A method for automatic evaluation of machine translation","volume-title":"Proc. 40th Annu. Meeting Assoc. Comput. Linguistics","author":"Papineni"},{"key":"ref34","first-page":"65","article-title":"METEOR: An automatic metric for MT evaluation with improved correlation with human judgments","volume-title":"Proc. acl workshop intrinsic extrinsic Eval. measures Mach. Transl. and\/or summarization","author":"Banerjee"},{"key":"ref35","first-page":"74","article-title":"ROUGE: A package for automatic evaluation of summaries","volume-title":"Text Summarization Branches Out","author":"Lin","year":"2004"},{"key":"ref36","article-title":"BERTScore: Evaluating text generation with BERT","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Zhang"},{"issue":"8","key":"ref37","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i17.29832"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.381"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1138"},{"key":"ref41","article-title":"Multiple-attribute text rewriting","volume-title":"Proc. 7th Int. Conf. Learn. Represent.","author":"Lample"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1601"},{"key":"ref43","article-title":"Formality style transfer with hybrid textual annotations","author":"Xu","year":"2019","journal-title":"arXiv:1903.06353"},{"key":"ref44","article-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling","volume-title":"Proc. NIPS Workshop Deep Learn.","author":"Chung"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1511.08458"},{"key":"ref46","first-page":"6830","article-title":"Style transfer from nonparallel text by cross-alignment","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Shen"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/526"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-1090"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1320"},{"key":"ref50","article-title":"Efficient reinforcement learning for unsupervised controlled text generation","author":"Upadhyay","year":"2022","journal-title":"arXiv:2204.07696"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref52","volume-title":"Improving language understanding by generative pretraining","author":"Radford","year":"2018"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref54","article-title":"Finetuned language models are zero-shot learners","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Lee"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.32"},{"key":"ref56","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Lee"},{"key":"ref57","first-page":"2790","article-title":"Parameter-efficient transfer learning for NLP","volume-title":"Proc. 36th Int. Conf. Mach. Learn.","volume":"97","author":"Houlsby"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acllong.353"},{"key":"ref59","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive NLP tasks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Lewis"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1139"},{"key":"ref61","first-page":"37309","article-title":"Deep bidirectional language-knowledge graph pretraining","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Yasunaga"},{"key":"ref62","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. Annu. Conf. Neural Inf. Process. Syst.","author":"Brown"},{"key":"ref63","article-title":"LaMDA: Language models for dialog applications","author":"Thoppilan","year":"2022","journal-title":"arXiv:2201.08239"},{"key":"ref64","author":"Wang","year":"2021","journal-title":"Mesh-Transformer-JAX: Model-Parallel Implementation of Transformer Language Model With JAX"},{"key":"ref65","author":"Black","year":"2021","journal-title":"GPT-Neo: Large Scale Autoregressive Language Modeling With Mesh-Tensorflow"},{"key":"ref66","article-title":"BART: Denoising sequence-tosequence pre-training for natural language generation, translation, and comprehension","author":"Lewis","year":"2019","journal-title":"arXiv:1910.13461"},{"key":"ref67","article-title":"Llama 2: Open foundation and fine-tuned chat models","author":"Touvron","year":"2023","journal-title":"arXiv:2307.09288"},{"key":"ref68","volume-title":"ChatGPT","year":"2021"},{"key":"ref69","article-title":"Are large language models actually good at text style transfer?","author":"Mukherjee","year":"2024","journal-title":"arXiv:2406.05885"},{"key":"ref70","first-page":"10755","article-title":"Falcon-40b: An open large language model with state-of-the-art performance","volume-title":"Proc. Findings Assoc. Comput. Linguistics, ACL","author":"Almazrouei"},{"key":"ref71","article-title":"LLaMA: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"arXiv:2302.13971"},{"key":"ref72","article-title":"Mistral 7B","author":"Jiang","year":"2023","journal-title":"arXiv:2310.06825"},{"key":"ref73","article-title":"BLOOM: A 176B-parameter open-access multilingual language model","author":"Le Scao","year":"2023","journal-title":"arXiv:2211.05100"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"ref75","article-title":"OPT: Open pre-trained transformer language models","author":"Zhang","year":"2022","journal-title":"arXiv:2205.01068"},{"key":"ref76","article-title":"Zephyr: Direct distillation of LM alignment","author":"Tunstall","year":"2023","journal-title":"arXiv:2310.16944"},{"key":"ref77","first-page":"15285","article-title":"Step-by-step: Controlling arbitrary style in text with large language models","volume-title":"Proc. Joint Int. Conf. Comput. Linguistics, Lang. Resour. Eval. (LREC-COLING)","author":"Liu"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.naacl-srw.21"},{"key":"ref79","article-title":"PaLM 2 technical report","volume-title":"arXiv:2305.10403","author":"Anil","year":"2023"},{"key":"ref80","article-title":"Unsupervised text style transfer via LLMs and attention masking with multi-way interactions","author":"Pan","year":"2024","journal-title":"arXiv:2402.13647"},{"issue":"1","key":"ref81","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"J. Mach. Learn. Res."},{"key":"ref82","first-page":"46","article-title":"Bubble up\u2014A fine-tuning approach for style transfer to community-specific subreddit language","volume-title":"Proc. 3rd Workshop Comput. Linguistics Political Social Sci.","author":"Zarcone"},{"key":"ref83","doi-asserted-by":"publisher","DOI":"10.1145\/3678179"},{"key":"ref84","first-page":"76","article-title":"Emulating author style: A feasibility study of prompt-enabled text stylization with off-the-shelf LLMs","volume-title":"Proc. 1st Workshop Personalization Generative AI Syst.","author":"Bhandarkar"},{"key":"ref85","article-title":"GPT-4 technical report","volume-title":"arXiv:2303.08774","year":"2023"},{"key":"ref86","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.990"},{"key":"ref87","first-page":"782","article-title":"Style vectors for steering generative large language models","volume-title":"Proc. Findings Assoc. Comput. Linguistics, EACL","author":"Konen"},{"key":"ref88","author":"Taori","year":"2023","journal-title":"Stanford Alpaca: An Instruction- Following LLaMA Model"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.745"},{"key":"ref90","article-title":"The Llama 3 herd of models","author":"Dubey","year":"2024","journal-title":"arXiv:2407.21783"},{"key":"ref91","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.222"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-long.244"},{"key":"ref93","doi-asserted-by":"publisher","DOI":"10.3233\/AISE240025"},{"key":"ref94","first-page":"123","article-title":"Style knowledge graph: Augmenting text style transfer with knowledge graphs","volume-title":"Proc. Int. Workshop Generative AI Knowl. Graphs@ GenAIK-COLING","author":"Toshevska"},{"key":"ref95","article-title":"Scaling instruction-finetuned language models","author":"Won Chung","year":"2022","journal-title":"arXiv:2210.11416"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.345"},{"issue":"3","key":"ref97","first-page":"3","article-title":"Phi-2: The surprising power of small language models","volume":"1","author":"Javaheripi","year":"2023","journal-title":"Microsoft Res. Blog"},{"key":"ref98","volume-title":"Artificial Intelligence a Modern Approach","author":"Russell","year":"2010"},{"key":"ref99","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref100","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017","journal-title":"arXiv:1707.06347"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.238"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-long.854"},{"key":"ref103","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.ijcnlp-main.9"},{"key":"ref104","article-title":"LMStyle benchmark: Evaluating text style transfer for chatbots","author":"Chen","year":"2024","journal-title":"arXiv:2403.08943"},{"key":"ref105","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.104"},{"key":"ref106","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proc. Conf. North Amer. Chapter Assoc. Comput. Linguistics, Hum. Lang. Technol.","author":"Devlin"},{"issue":"240","key":"ref107","first-page":"1","article-title":"PaLM: Scaling language modeling with pathways","volume":"24","author":"Chowdhery","year":"2022","journal-title":"J. Mach. Learn. Res."},{"key":"ref108","article-title":"Gemini:Afamily of highly capable multimodal models","author":"Team","year":"2023","journal-title":"arXiv:2312.11805"},{"key":"ref109","doi-asserted-by":"publisher","DOI":"10.1145\/3411763.3451760"},{"key":"ref110","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.295"},{"key":"ref111","article-title":"Least-to-most prompting enables complex reasoning in large language models","volume-title":"Proc. 11th Int. Conf. Learn. Represent.","author":"Zhou"},{"key":"ref112","article-title":"A systematic survey of prompt engineering in large language models: Techniques and applications","author":"Sahoo","year":"2024","journal-title":"arXiv:2402.07927"},{"key":"ref113","article-title":"Asurvey of prompt engineering methods in large language models for different NLP tasks","author":"Vatsal","year":"2024","journal-title":"arXiv:2407.12994"},{"key":"ref114","first-page":"22199","article-title":"Large language models are zero-shot reasoners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Kojima"},{"key":"ref115","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.147"},{"key":"ref116","first-page":"1","article-title":"Skeletonof- thought: Prompting LLMs for efficient parallel generation","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Ning"},{"key":"ref117","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.212"},{"key":"ref118","article-title":"Parameterefficient fine-tuning methods for pretrained language models: A critical review and assessment","author":"Xu","year":"2023","journal-title":"arXiv:2312.12148"},{"key":"ref119","article-title":"Parameterefficient fine-tuning for large models: A comprehensive survey","author":"Han","year":"2024","journal-title":"arXiv:2403.14608"},{"issue":"120","key":"ref120","first-page":"1","article-title":"Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity","volume":"23","author":"Fedus","year":"2021","journal-title":"J. Mach. Learn. Res."},{"key":"ref121","article-title":"Scaling language models: Methods, analysis&insights from training gopher","author":"Rae","year":"2021","journal-title":"arXiv:2112.11446"},{"key":"ref122","article-title":"Instruction tuning for large language models: A survey","author":"Zhang","year":"2023","journal-title":"arXiv:2308.10792"},{"key":"ref123","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Ouyang"},{"key":"ref124","article-title":"Multitask prompted training enables zero-shot task generalization","volume-title":"Proc. 10th Int. Conf. Learn. Represent.","author":"Sanh"},{"key":"ref125","first-page":"4302","article-title":"Deep reinforcement learning from human preferences","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Christiano"},{"key":"ref126","article-title":"A survey of reinforcement learning from human feedback","author":"Kaufmann","year":"2023","journal-title":"arXiv:2312.14925"},{"key":"ref127","first-page":"53728","article-title":"Direct preference optimization:Your language model is secretly a reward model","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Rafailov"},{"key":"ref128","article-title":"Retrieval-augmented generation for large language models: A survey","author":"Gao","year":"2023","journal-title":"arXiv:2312.10997"},{"key":"ref129","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.322"},{"key":"ref130","doi-asserted-by":"publisher","DOI":"10.1145\/3589335.3648298"},{"key":"ref131","article-title":"Take a step back: Evoking reasoning via abstraction in large language models","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Zheng"},{"key":"ref132","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.99"},{"key":"ref133","article-title":"KnowledGPT: Enhancing large language models with retrieval and storage access on knowledge bases","author":"Wang","year":"2023","journal-title":"arXiv:2308.11761"},{"key":"ref134","doi-asserted-by":"publisher","DOI":"10.5121\/ijnlc.2024.13103"},{"key":"ref135","article-title":"From classification to generation: Insights into crosslingual retrieval augmented ICL","volume-title":"Proc. NeurIPS Workshop Instruct. Tuning Instruct. Following","author":"Li"},{"key":"ref136","article-title":"A survey on knowledge-enhanced pre-trained language models","author":"Zhen","year":"2022","journal-title":"arXiv:2212.13428"},{"key":"ref137","article-title":"Language models are open knowledge graphs","author":"Wang","year":"2020","journal-title":"arXiv:2010.11967"},{"key":"ref138","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.437"},{"key":"ref139","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.34"},{"key":"ref140","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1005"},{"key":"ref141","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.551"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/10915631.pdf?arnumber=10915631","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,17]],"date-time":"2025-03-17T17:48:47Z","timestamp":1742233727000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10915631\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":141,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3548967","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}