{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,2]],"date-time":"2026-05-02T09:29:57Z","timestamp":1777714197717,"version":"3.51.4"},"reference-count":128,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100005145","name":"Basic Research Projects in Liaoning Provincial Undergraduate Universities","doi-asserted-by":"publisher","award":["SYLUGXTD07"],"award-info":[{"award-number":["SYLUGXTD07"]}],"id":[{"id":"10.13039\/501100005145","id-type":"DOI","asserted-by":"publisher"}]},{"name":"General Young Talents Project for Scientific Research Grant of the Educational Department of Liaoning Province","award":["LJKQR20222512"],"award-info":[{"award-number":["LJKQR20222512"]}]},{"name":"Research Support Program for Inviting High-Level Talents Grant of Shenyang Ligong University","award":["1010147001004"],"award-info":[{"award-number":["1010147001004"]}]},{"DOI":"10.13039\/501100004543","name":"China Scholarship Council","doi-asserted-by":"publisher","award":["202208210120"],"award-info":[{"award-number":["202208210120"]}],"id":[{"id":"10.13039\/501100004543","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/access.2024.3487352","type":"journal-article","created":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T17:39:46Z","timestamp":1730137186000},"page":"179353-179382","source":"Crossref","is-referenced-by-count":29,"title":["Recent Advances in Interactive Machine Translation With Large Language Models"],"prefix":"10.1109","volume":"12","author":[{"given":"Yanshu","family":"Wang","sequence":"first","affiliation":[{"name":"Art and Design College, Shenyang Ligong University, Shenyang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9813-217X","authenticated-orcid":false,"given":"Jinyi","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Information Science and Engineering, Shenyang Ligong University, Shenyang, China"}]},{"given":"Tianrong","family":"Shi","sequence":"additional","affiliation":[{"name":"School of Information Science and Engineering, Shenyang Ligong University, Shenyang, China"}]},{"given":"Dashuai","family":"Deng","sequence":"additional","affiliation":[{"name":"School of Information Science and Engineering, Shenyang Ligong University, Shenyang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0981-6677","authenticated-orcid":false,"given":"Ye","family":"Tian","sequence":"additional","affiliation":[{"name":"Zhuzhou CRRC Times Electric Company Ltd., Zhuzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4106-1099","authenticated-orcid":false,"given":"Tadahiro","family":"Matsumoto","sequence":"additional","affiliation":[{"name":"Faculty of Engineering, Gifu University, Gifu, Japan"}]}],"member":"263","reference":[{"key":"ref1","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","volume-title":"Proc. Annu. Conf. Neural Inf. Process. Syst.","author":"Sutskever"},{"key":"ref2","first-page":"1","article-title":"Neural machine translation by jointly learning to align and translate","volume-title":"Proc. 3rd Int. Conf. Learn. Represent. (ICLR)","author":"Bahdanau"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/d15-1166"},{"key":"ref4","first-page":"5998","article-title":"Attention is all you need","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Vaswani"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/s10590-014-9167-7"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/s10590-015-9175-2"},{"key":"ref7","article-title":"GPT- 4 technical report","volume-title":"arXiv.2303.08774","author":"Achiam","year":"2023"},{"key":"ref8","first-page":"107","article-title":"Neural interactive translation prediction","volume-title":"Proc. 12nd Conferences Assoc. Mach. Transl. Americas,MTResearchers\u2019 Track","author":"Knowles"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/d19-3018"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/p19-1175"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.144"},{"key":"ref12","article-title":"How good are GPT models at machine translation? A comprehensive evaluation","author":"Hendy","year":"2023","journal-title":"arXiv:2302.09210"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-naacl.176"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.wmt-1.40"},{"key":"ref15","author":"Jiao","year":"2023","journal-title":"Is ChatGPT a Good Translator? Yes With GPT-4 As the Engine"},{"key":"ref16","first-page":"3111","article-title":"Distributed representations of words and phrases and their compositionality","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"2","author":"Mikolov"},{"key":"ref17","first-page":"1","article-title":"Language models are few-shot learners","author":"Brown","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"240","key":"ref18","first-page":"1","article-title":"PaLM: Scaling language modeling with pathways","volume":"24","author":"Chowdhery","year":"2023","journal-title":"J. Mach. Learn. Res."},{"key":"ref19","article-title":"BLOOM: A 176b-parameter open-access multilingual language model","author":"Scao","year":"2022","journal-title":"arXiv.2211.05100"},{"key":"ref20","article-title":"OPT: Open pre-trained transformer language models","author":"Zhang","year":"2022","journal-title":"arXiv:2205.01068"},{"key":"ref21","article-title":"LLaMA: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"arXiv:2302.13971"},{"key":"ref22","article-title":"Llama 2: Open foundation and fine-tuned chat models","author":"Touvron","year":"2023","journal-title":"arXiv:2307.09288"},{"key":"ref23","article-title":"No language left behind: Scaling human-centered machine translation","author":"Team","year":"2022","journal-title":"arXiv:2207.04672"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-33-6162-1_1"},{"issue":"990","key":"ref25","article-title":"Thai-chinese neural machine translation method based on dependency distance penalty","volume":"55","author":"Zhang","year":"2022","journal-title":"Commun. Technol."},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.616"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.564"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.859"},{"key":"ref29","first-page":"41092","article-title":"Prompting large language model for machine translation: A case study","volume-title":"Proc. Int. Conf. Mach. Learn.","volume":"202","author":"Zhang"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i11.26585"},{"key":"ref31","first-page":"227","article-title":"Adaptive machine translation with large language models","volume-title":"Proc. 24th Annu. Conf. Eur. Assoc. Mach. Transl.","author":"Moslem"},{"key":"ref32","first-page":"10867","article-title":"The unreasonable effectiveness of few-shot learning for machine translation","volume-title":"Proc. Int. Conf. Mach. Learn.","volume":"202","author":"Garcia"},{"key":"ref33","article-title":"Dictionary-based phrase-level prompting of large language models for machine translation","author":"Ghazvininejad","year":"2023","journal-title":"arXiv:2302.07856"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-short.126"},{"key":"ref35","article-title":"Instruction position matters in sequence generation with large language models","author":"Liu","year":"2023","journal-title":"arXiv:2308.12097"},{"key":"ref36","article-title":"Improving translation faithfulness of large language models via augmenting instructions","author":"Chen","year":"2023","journal-title":"arXiv:2308.12674"},{"key":"ref37","article-title":"Neural machine translation models can learn to be few-shot learners","author":"Reinauer","year":"2023","journal-title":"arXiv:2309.08590"},{"key":"ref38","first-page":"482","article-title":"Towards effective disambiguation for machine translation with large language models","volume-title":"Proc. 8th Conf. Mach. Transl.","author":"Iyer"},{"key":"ref39","first-page":"11127","article-title":"Steering large language models for machine translation with finetuning and in-context learning","volume-title":"Proc. Findings Assoc. Comput. Linguistics","author":"Alves"},{"key":"ref40","article-title":"Dissecting in-context learning of translations in GPTs","author":"Raunak","year":"2023","journal-title":"arXiv:2310.15987"},{"key":"ref41","article-title":"Narrowing the gap between zero- and few-shot machine translation by matching styles","author":"Tan","year":"2023","journal-title":"arXiv:2311.02310"},{"key":"ref42","article-title":"Anti-LM decoding for zero-shot incontext machine translation","author":"Sia","year":"2023","journal-title":"arXiv:2311.08324"},{"key":"ref43","first-page":"8616","article-title":"MT2: Towards a multi-task machine translation model with translationspecific in-context learning","volume-title":"Proc. Conf. Empirical Methods Natural Lang. Process.","author":"Li"},{"key":"ref44","article-title":"Self-alignment with instruction backtranslation","author":"Li","year":"2023","journal-title":"arXiv:2308.06259"},{"key":"ref45","article-title":"Take a step back: Evoking reasoning via abstraction in large language models","author":"Zheng","year":"2023","journal-title":"arXiv:2310.06117"},{"key":"ref46","article-title":"Exploring human-like translation strategy with large language models","author":"He","year":"2023","journal-title":"arXiv:2305.04118"},{"key":"ref47","first-page":"5622","article-title":"Towards making the most of ChatGPT for machine translation","volume-title":"Proc. Findings Assoc. Comput. Linguistics","author":"Peng"},{"key":"ref48","article-title":"Chain-ofdictionary prompting elicits translation in large language models","author":"Lu","year":"2023","journal-title":"arXiv:2305.06575"},{"key":"ref49","article-title":"Encouraging divergent thinking in large language models through multi-agent debate","author":"Liang","year":"2023","journal-title":"arXiv:2305.19118"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.289"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.320"},{"key":"ref52","article-title":"Reasoning chain based adversarial attack for multi-hop question answering","author":"Ding","year":"2021","journal-title":"arXiv:2112.09658"},{"key":"ref53","first-page":"24824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Wei"},{"key":"ref54","article-title":"Cross-lingual supervision improves large language models pre-training","author":"Schioppa","year":"2023","journal-title":"arXiv:2305.11778"},{"key":"ref55","article-title":"PolyLM: An open source polyglot large language model","author":"Wei","year":"2023","journal-title":"arXiv:2307.06018"},{"key":"ref56","article-title":"OpenBA: An open-sourced 15B bilingual asymmetric seq2seq model pre-trained from scratch","author":"Li","year":"2023","journal-title":"arXiv:2309.10706"},{"key":"ref57","article-title":"Gemini: A family of highly capable multimodal models","author":"Anil","year":"2023","journal-title":"arXiv.2312.11805"},{"key":"ref58","article-title":"The falcon series of open language models","author":"Almazrouei","year":"2023","journal-title":"arXiv:2311.16867"},{"key":"ref59","first-page":"1","article-title":"LoRa: Low-rank adaptation of large language models","volume-title":"Proc. 10th Int. Conf. Learn. Represent. (ICLR)","author":"Hu"},{"key":"ref60","article-title":"Higher layers need more LoRa experts","author":"Gao","year":"2024","journal-title":"arXiv:2402.08562"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.loresmt-1.1"},{"key":"ref62","article-title":"A paradigm shift in machine translation: Boosting translation performance of large language models","author":"Xu","year":"2023","journal-title":"arXiv:2309.11674"},{"key":"ref63","first-page":"15009","article-title":"ParroT: Translating during chat using large language models tuned with human translation and feedback","volume-title":"Proc. Findings Assoc. Comput. Linguistics","author":"Jiao"},{"key":"ref64","article-title":"Eliciting the translation ability of large language models via multilingual finetuning with translation instructions","author":"Li","year":"2023","journal-title":"arXiv:2305.15083"},{"key":"ref65","article-title":"BigTranslate: Augmenting large language models with multilingual translation capability over 100 languages","author":"Yang","year":"2023","journal-title":"arXiv.2305.18098"},{"key":"ref66","article-title":"BayLing: Bridging cross-lingual alignment and instruction following through interactive translation for large language models","author":"Zhang","year":"2023","journal-title":"arXiv:2306.10968"},{"key":"ref67","article-title":"TIM: Teaching large language models to translate with comparison","author":"Zeng","year":"2023","journal-title":"arXiv:2307. 04408"},{"key":"ref68","article-title":"Extrapolating large language models to non-english by aligning languages","author":"Zhu","year":"2023","journal-title":"arXiv:2308.04948"},{"key":"ref69","first-page":"14","article-title":"Domainspecific text generation for machine translation","volume-title":"Proc. 15th Biennial Conf. Assoc. Mach. Transl. Americas","author":"Moslem"},{"key":"ref70","article-title":"Fine-tuning large language models for adaptive machine translation","author":"Moslem","year":"2023","journal-title":"arXiv:2312.12740"},{"key":"ref71","article-title":"Adapting large language models for document-level machine translation","author":"Wu","year":"2024","journal-title":"arXiv:2401.06468"},{"key":"ref72","article-title":"Contrastive preference optimization: Pushing the boundaries of LLM performance in machine translation","author":"Xu","year":"2024","journal-title":"arXiv:2401.08417"},{"key":"ref73","first-page":"17359","article-title":"Locating and editing factual associations in GPT","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Meng"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-44693-1_33"},{"key":"ref75","article-title":"New trends in machine translation using large language models: Case examples with ChatGPT","author":"Lyu","year":"2023","journal-title":"arXiv.2305.01181"},{"key":"ref76","first-page":"166","article-title":"Exploring prompt engineering with GPT language models for document-level machine translation: Insights and findings","volume-title":"Proc. 8th Conf. Mach. Transl.","author":"Wu"},{"key":"ref77","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.599"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2023.101566"},{"key":"ref79","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i11.26490"},{"key":"ref80","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-86383-8_26"},{"key":"ref81","article-title":"GPTScore: Evaluate as you desire","author":"Fu","year":"2023","journal-title":"arXiv:2302.04166"},{"key":"ref82","first-page":"193","article-title":"Large language models are state-ofthe- art evaluators of translation quality","volume-title":"Proc. 24th Annu. Conf. Eur. Assoc. Mach. Transl.","author":"Kocmi"},{"key":"ref83","article-title":"Error analysis prompting enables human-like translation evaluation in large language models: A case study on ChatGPT","author":"Lu","year":"2023","journal-title":"arXiv.2303.13809"},{"key":"ref84","first-page":"5967","article-title":"INSTRUCTSCORE: Towards explainable text generation evaluation with automatic feedback","volume-title":"Proc. Conf. Empirical Methods Natural Lang. Process.","author":"Xu"},{"key":"ref85","article-title":"Towards explainable evaluation metrics for machine translation","author":"Leiter","year":"2023","journal-title":"arXiv:2306.13041"},{"key":"ref86","first-page":"1066","article-title":"The devil is in the errors: Leveraging large language models for fine-grained machine translation evaluation","volume-title":"Proc. 8th Conf. Mach. Transl.","author":"Fernandes"},{"key":"ref87","article-title":"Towards multiple references era\u2014Addressing data leakage and limited reference diversity in NLG evaluation","author":"Zeng","year":"2023","journal-title":"arXiv.2308.03131"},{"key":"ref88","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.211"},{"key":"ref89","first-page":"12009","article-title":"Leveraging GPT-4 for automatic translation post-editing","volume-title":"Proc. Findings Assoc. Comput. Linguistics","author":"Raunak"},{"key":"ref90","article-title":"Iterative translation refinement with large language models","author":"Chen","year":"2023","journal-title":"arXiv:2306.03856"},{"key":"ref91","article-title":"Contextual refinement of translations: Large language models for sentence and document-level post-editing","author":"Koneru","year":"2023","journal-title":"arXiv:2310.14855"},{"key":"ref92","article-title":"SCALE: Synergized collaboration of asymmetric language translation engines","author":"Cheng","year":"2023","journal-title":"arXiv:2309.17061"},{"key":"ref93","first-page":"902","article-title":"Domain terminology integration into machine translation: Leveraging large language models","volume-title":"Proc. 8th Conf. Mach. Transl.","author":"Moslem"},{"key":"ref94","article-title":"Beyond the chat: Executable and verifiable text-editing with LLMs","author":"Laban","year":"2023","journal-title":"arXiv:2309.15337"},{"key":"ref95","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-demo.19"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-short.122"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.524"},{"key":"ref98","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/587"},{"key":"ref99","first-page":"1","article-title":"The reasonableness behind unreasonable translation capability of large language models","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Fu"},{"key":"ref100","article-title":"Improving machine translation with large language models: A preliminary study with cooperative decoding","author":"Zeng","year":"2023","journal-title":"arXiv:2311.02851"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-naacl.35"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.275"},{"key":"ref103","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.ijcnlp-main.45"},{"key":"ref104","first-page":"16646","article-title":"Documentlevel machine translation with large language models","volume-title":"Proc. Conf. Empirical Methods Natural Lang. Process.","author":"Wang"},{"key":"ref105","article-title":"How to design translation prompts for ChatGPT: An empirical study","author":"Gao","year":"2023","journal-title":"arXiv:2304.02182"},{"key":"ref106","first-page":"157","article-title":"Investigating the translation performance of a large multilingual language model: The case of BLOOM","volume-title":"Proc. 24th Annu. Conf. Eur. Assoc. Mach. Transl.","author":"Bawden"},{"key":"ref107","first-page":"419","article-title":"Large language models effectively leverage document-level context for literary translation, but critical errors persist","volume-title":"Proc. 8th Conf. Mach. Transl.","author":"Karpinska"},{"key":"ref108","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-short.90"},{"key":"ref109","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-44693-1_30"},{"key":"ref110","first-page":"10176","article-title":"Cceval:Arepresentative evaluation benchmark for the chinese-centric multilingual machine translation","volume-title":"Proc. Findings Assoc. Comput. Linguistics","author":"Lou"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.26615\/issn.2683-0078.2023_008"},{"key":"ref112","article-title":"Textbooks are all you need II: Phi-1.5 technical report","volume-title":"arXiv:2309.05463","author":"Li","year":"2023"},{"key":"ref113","article-title":"Knowledge sanitization of large language models","author":"Ishibashi","year":"2023","journal-title":"arXiv:2309.11852"},{"key":"ref114","first-page":"311","article-title":"BLEU: A method for automatic evaluation of machine translation","volume-title":"Proc. 40th Annu. Meeting Assoc. Comput. Linguistics","author":"Papineni"},{"key":"ref115","first-page":"578","article-title":"COMET- 22: Unbabel-ist 2022 submission for the metrics shared task","volume-title":"Proc. 7th Conf. Mach. Transl.","author":"Rei"},{"key":"ref116","first-page":"569","article-title":"MaTESe: Machine translation evaluation as a sequence tagging problem","volume-title":"Proc. 7th Conf. Mach. Transl.","author":"Perrella"},{"key":"ref117","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.704"},{"key":"ref118","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/w15-3049"},{"key":"ref119","first-page":"65","article-title":"METEOR: An automatic metric for MT evaluation with improved correlation with human judgments","volume-title":"Proc. Workshop Intrinsic Extrinsic Eval. Measures Mach. Transl. Summarization@ACL","author":"Banerjee"},{"key":"ref120","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr.2015.7299087"},{"key":"ref121","first-page":"1","article-title":"BERTScore: Evaluating text generation with BERT","volume-title":"Proc. 8th Int. Conf. Learn. Represent.","author":"Zhang"},{"key":"ref122","first-page":"27263","article-title":"BARTScore: Evaluating generated text as text generation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Yuan"},{"key":"ref123","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.8"},{"key":"ref124","article-title":"SEScore2: Retrieval augmented pretraining for text generation evaluation","author":"Xu","year":"2022","journal-title":"arXiv.2212.09305"},{"key":"ref125","first-page":"46","article-title":"Results of WMT22 metrics shared task: Stop using BLEU\u2014Neural metrics are better and more robust","volume-title":"Proc. 7th Conf. Mach. Transl.","author":"Freitag"},{"key":"ref126","first-page":"4188","article-title":"Framing image description as a ranking task: Data, models and evaluation metrics (extended abstract)","volume-title":"Proc. 24th Int. Joint Conf. Artif. Intell.","author":"Hodosh"},{"key":"ref127","first-page":"1552","article-title":"Phrase-based statistical language generation using graphical models and active learning","volume-title":"Proc. 48th Annu. Meeting Assoc. Comput. Linguistics","author":"Mairesse"},{"key":"ref128","first-page":"1","article-title":"CommonGen: A constrained text generation challenge for generative commonsense reasoning","volume-title":"Proc. Conf. Automated Knowl. Base Construction","author":"Lin"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10380310\/10737082.pdf?arnumber=10737082","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,24]],"date-time":"2025-04-24T17:06:47Z","timestamp":1745514407000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10737082\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":128,"URL":"https:\/\/doi.org\/10.1109\/access.2024.3487352","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]}}}