{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T16:19:34Z","timestamp":1777911574193,"version":"3.51.4"},"reference-count":138,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["82404831"],"award-info":[{"award-number":["82404831"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["2632024RWPY04"],"award-info":[{"award-number":["2632024RWPY04"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["2632024PY01"],"award-info":[{"award-number":["2632024PY01"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3552256","type":"journal-article","created":{"date-parts":[[2025,3,18]],"date-time":"2025-03-18T17:40:38Z","timestamp":1742319638000},"page":"51110-51129","source":"Crossref","is-referenced-by-count":8,"title":["A Survey of Large Language Model for Drug Research and Development"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-9967-4316","authenticated-orcid":false,"given":"Huijie","family":"Guo","sequence":"first","affiliation":[{"name":"Information Construction Management Office, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3081-1694","authenticated-orcid":false,"given":"Xudong","family":"Xing","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Natural Medicines, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Yongjie","family":"Zhou","sequence":"additional","affiliation":[{"name":"College of Science, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Wenjiao","family":"Jiang","sequence":"additional","affiliation":[{"name":"Center for Excellence in Molecular Cell Science, Chinese Academy of Sciences, Shanghai, China"}]},{"given":"Xiaoyi","family":"Chen","sequence":"additional","affiliation":[{"name":"College of Science, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Ting","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Natural Medicines, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Zixuan","family":"Jiang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Natural Medicines, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Yibing","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Natural Medicines, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Junyan","family":"Hou","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Natural Medicines, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Yukun","family":"Jiang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Natural Medicines, China Pharmaceutical University, Nanjing, Jiangsu, China"}]},{"given":"Jianzhen","family":"Xu","sequence":"additional","affiliation":[{"name":"College of Science, China Pharmaceutical University, Nanjing, Jiangsu, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i10.28948"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3112535"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1038\/s41592-021-01252-x"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.61969\/jai.1311271"},{"key":"ref5","article-title":"Layer normalization","author":"Ba","year":"2016","journal-title":"arXiv:1607.06450"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1098\/rstb.2019.0307"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/d19-1371"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1093\/jamiaopen\/ooac043"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.inlg-main.25"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btac020"},{"key":"ref11","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Brown"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.52591\/lxai202106257"},{"key":"ref13","article-title":"XTrimoPGLM: Unified 100B-scale pre-trained transformer for deciphering the language of protein","author":"Chen","year":"2024","journal-title":"arXiv:2401.06199"},{"key":"ref14","article-title":"Interpretable RNA foundation model from unannotated data for highly accurate RNA structure and function predictions","author":"Chen","year":"2022","journal-title":"arXiv:2204.00300"},{"key":"ref15","article-title":"BianQue: Balancing the questioning and suggestion ability of health LLMs with multi-turn health conversations polished by ChatGPT","author":"Chen","year":"2023","journal-title":"arXiv:2310.15896"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.83"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_7"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.drudis.2021.08.010"},{"key":"ref19","article-title":"PaLM: Scaling language modeling with pathways","author":"Chowdhery","year":"2022","journal-title":"J. Mach. Learn. Res."},{"issue":"70","key":"ref20","first-page":"1","article-title":"Scaling instruction-finetuned language models","volume":"25","author":"Chung","year":"2022","journal-title":"J. Mach. Learn. Res."},{"key":"ref21","article-title":"Transformer-XL: Attentive language models beyond a fixed-length context","author":"Dai","year":"2019","journal-title":"arXiv:1901.02860"},{"key":"ref22","article-title":"The nucleotide transformer: Building and evaluating robust foundation models for human genomics","author":"Dalla-Torre","year":"2023","journal-title":"bioRxiv"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/3448734.3450772"},{"key":"ref24","article-title":"QLoRA: Efficient finetuning of quantized LLMs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Dettmers"},{"key":"ref25","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv:1810.04805"},{"issue":"3","key":"ref26","doi-asserted-by":"crossref","first-page":"220","DOI":"10.1038\/s42256-023-00626-4","article-title":"Parameter-efficient fine-tuning of large-scale pre-trained language models","volume":"5","author":"Ding","year":"2023","journal-title":"Nature Mach. Intell."},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/BIBM52615.2021.9669588"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2010.11929"},{"key":"ref29","article-title":"SynerGPT: In-context learning for personalized drug synergy prediction and drug design","author":"Edwards","year":"2023","journal-title":"arXiv:2307.11694"},{"issue":"1","key":"ref30","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s12967-017-1285-6","article-title":"In silico prediction of novel therapeutic targets using gene\u2013disease association data","volume":"15","author":"Ferrero","year":"2017","journal-title":"J. Transl. Med."},{"issue":"1","key":"ref31","doi-asserted-by":"crossref","first-page":"4348","DOI":"10.1038\/s41467-022-32007-7","article-title":"ProtGPT2 is a deep unsupervised language model for protein design","volume":"13","author":"Ferruz","year":"2022","journal-title":"Nature Commun."},{"key":"ref32","article-title":"GENA-LM: A family of open-source foundational models for long DNA sequences","author":"Fishman","year":"2023","journal-title":"bioRxiv"},{"key":"ref33","article-title":"Ziya2: Data-centric learning is all LLMs need","author":"Gan","year":"2023","journal-title":"arXiv:2311.03301"},{"key":"ref34","article-title":"Ahead-of-time P-tuning","author":"Gavrilov","year":"2023","journal-title":"arXiv:2305.10835"},{"key":"ref35","first-page":"1","article-title":"Modern methods of drug discovery: An introduction","author":"Giersiefen","year":"2003","journal-title":"Modern Methods of Drug Discovery"},{"key":"ref36","article-title":"It\u2019s not Greek to mBERT: Inducing word-level translations from multilingual BERT","author":"Gonen","year":"2020","journal-title":"arXiv:2010.08275"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.maiworkshop-1.10"},{"key":"ref38","article-title":"Aloe: A family of fine-tuned open healthcare LLMs","author":"Kumar Gururajan","year":"2024","journal-title":"arXiv:2405.01886"},{"key":"ref39","article-title":"Conformer: Convolution-augmented transformer for speech recognition","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Han"},{"issue":"1","key":"ref40","doi-asserted-by":"crossref","DOI":"10.1186\/s13321-016-0160-4","article-title":"Jmol SMILES and jmol SMARTS: Specifications and applications","volume":"8","author":"Hanson","year":"2016","journal-title":"J. Cheminformatics"},{"issue":"3","key":"ref41","doi-asserted-by":"crossref","first-page":"225","DOI":"10.1007\/s12192-023-01340-1","article-title":"What artificial intelligence knows about 70 kDa heat shock proteins, and how we will face this ChatGPT era","volume":"28","author":"Heck","year":"2023","journal-title":"Cell Stress Chaperones"},{"key":"ref42","article-title":"Measuring massive multitask language understanding","author":"Hendrycks","year":"2020","journal-title":"arXiv:2009.03300"},{"key":"ref43","article-title":"The diminishing returns of masked language models to science","author":"Hong","year":"2022","journal-title":"arXiv:2205.11342"},{"key":"ref44","article-title":"LoRA: Low-rank adaptation of large language models","author":"Hu","year":"2021","journal-title":"arXiv:2106.09685"},{"key":"ref45","article-title":"Predicting FDA approvability of small-molecule drugs","author":"Huang","year":"2022","journal-title":"bioRxiv"},{"key":"ref46","article-title":"ClinicalBERT: Modeling clinical notes and predicting hospital readmission","author":"Huang","year":"2019","journal-title":"arXiv:1904.05342"},{"key":"ref47","article-title":"Language is not all you need: Aligning perception with language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Huang"},{"issue":"7","key":"ref48","doi-asserted-by":"crossref","first-page":"603","DOI":"10.1016\/j.tig.2021.04.001","article-title":"Genetics in drug discovery","volume":"37","author":"Hubaud","year":"2021","journal-title":"Trends Genet."},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1111\/j.1476-5381.2010.01127.x"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1088\/2632-2153\/ac3ffb"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1080\/14656566.2022.2161366"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.2174\/156802662124211104100228"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btab083"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1259"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btae075"},{"key":"ref56","article-title":"Gene set summarization using large language models","author":"Joachimiak","year":"2023","journal-title":"arXiv:2305.13338"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.aacl-main.83"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-022-13428-4"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.324"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1208\/s12248-021-00644-3"},{"key":"ref61","article-title":"MedGPT: Medical concept prediction from clinical narratives","author":"Kraljevic","year":"2021","journal-title":"arXiv:2107.03134"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1088\/2632-2153\/aba947"},{"key":"ref63","article-title":"ALBERT: A lite BERT for self-supervised learning of language representations","author":"Lan","year":"2019","journal-title":"arXiv:1909.11942"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbab005"},{"key":"ref65","article-title":"Beyond scale: The diversity coefficient as a data quality metric for variability in natural language data","author":"Miranda","year":"2023","journal-title":"arXiv:2306.13840"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.936"},{"key":"ref68","article-title":"Empowering molecule discovery for molecule-caption translation with large language models: A ChatGPT perspective","author":"Li","year":"2023","journal-title":"arXiv:2306.06615"},{"key":"ref69","article-title":"Assessing the performance of GPT-4 in the filed of osteoarthritis and orthopaedic case consultation","author":"Li","year":"2023","journal-title":"medRxiv"},{"key":"ref70","article-title":"CodonBERT: Large language models for mRNA design and optimization","author":"Li","year":"2023","journal-title":"bioRxiv"},{"key":"ref71","article-title":"Prefix-tuning: Optimizing continuous prompts for generation","author":"Lisa Li","year":"2021","journal-title":"arXiv:2101.00190"},{"key":"ref72","article-title":"DrugGPT: A GPT-based strategy for designing potential ligands targeting specific proteins","author":"Li","year":"2023","journal-title":"bioRxiv"},{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1126\/science.ade2574"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.2196\/48568"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.3390\/ijms242316761"},{"key":"ref76","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.1c01467"},{"key":"ref77","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbac409"},{"key":"ref78","article-title":"KOSMOS-2.5: A multimodal literate model","author":"Lv","year":"2023","journal-title":"arXiv:2309.11419"},{"key":"ref79","article-title":"ProGen: Language modeling for protein generation","author":"Madani","year":"2020","journal-title":"arXiv:2004.03497"},{"key":"ref80","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3075573"},{"key":"ref81","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3406714"},{"key":"ref82","doi-asserted-by":"publisher","DOI":"10.1038\/nrd.2016.136"},{"key":"ref83","article-title":"HyenaDNA: Long-range genomic sequence modeling at single nucleotide resolution","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Nguyen"},{"key":"ref84","doi-asserted-by":"publisher","DOI":"10.1016\/j.cels.2023.10.002"},{"key":"ref85","article-title":"Structured chemistry reasoning with large language models","author":"Ouyang","year":"2023","journal-title":"arXiv:2311.09656"},{"key":"ref86","first-page":"248","article-title":"MedMCQA: A large-scale multi-subject multi-choice dataset for medical domain question answering","volume-title":"Proc. Conf. Health, Inference, Learn.","author":"Pal"},{"key":"ref87","doi-asserted-by":"publisher","DOI":"10.3102\/1076998619872761"},{"key":"ref88","article-title":"Kosmos-2: Grounding multimodal large language models to the world","author":"Peng","year":"2023","journal-title":"arXiv:2306.14824"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.1111\/j.1365-2125.1988.tb03318.x"},{"key":"ref90","article-title":"Improving language understanding by generative pre-training","author":"Radford","year":"2018"},{"issue":"8","key":"ref91","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"ref92","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","author":"Raffel","year":"2019","journal-title":"J. Mach. Learn. Res."},{"key":"ref93","article-title":"Hierarchical text-conditional image generation with CLIP latents","author":"Ramesh","year":"2022","journal-title":"arXiv:2204.06125"},{"key":"ref94","article-title":"Evaluation of gpt and bert-based models on identifying proteinprotein interactions in biomedical text","author":"Rehana","year":"2023","journal-title":"arXiv:2303.17728"},{"key":"ref95","article-title":"Computer vision","author":"Rosenfeld","year":"1988","journal-title":"Science"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.1007\/s40273-021-01065-y"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/10.6.635"},{"key":"ref98","doi-asserted-by":"publisher","DOI":"10.1016\/S0006-3495(94)80782-9"},{"key":"ref99","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2021.104851"},{"key":"ref100","first-page":"6401","article-title":"Natural TTS synthesis by conditioning WaveNet on MEL spectrogram predictions of tacotron","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Shen"},{"key":"ref101","article-title":"Large language models as subpopulation representative models: A review","author":"Simmons","year":"2023","journal-title":"arXiv:2310.17888"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.3389\/fmed.2021.760762"},{"key":"ref103","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-024-03423-7"},{"key":"ref104","doi-asserted-by":"publisher","DOI":"10.1145\/3582688"},{"key":"ref105","article-title":"VL-BERT: Pre-training of generic visual-linguistic representations","author":"Su","year":"2019","journal-title":"arXiv:1908.08530"},{"key":"ref106","doi-asserted-by":"publisher","DOI":"10.1111\/cts.12577"},{"key":"ref107","article-title":"Parameter-efficient prompt tuning makes generalized and calibrated neural text retrievers","author":"Lam Tam","year":"2022","journal-title":"arXiv:2207.07087"},{"key":"ref108","article-title":"MedChatZH: A better medical adviser learns from better instructions","author":"Tan","year":"2023","journal-title":"arXiv:2309.01114"},{"key":"ref109","article-title":"ChiMed-GPT: A Chinese medical large language model with full training regime and better alignment to human preferences","author":"Tian","year":"2023","journal-title":"arXiv:2311.06025"},{"key":"ref110","article-title":"LLaMA: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"arXiv:2302.13971"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocae122"},{"key":"ref112","article-title":"One-shot learning for language modelling","author":"Ucar","year":"2020","journal-title":"arXiv:2007.09679"},{"key":"ref113","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref114","article-title":"SuperGLUE: A stickier benchmark for general-purpose language understanding systems","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref115","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5446"},{"key":"ref116","article-title":"ClinicalGPT: Large language models finetuned with diverse medical data and comprehensive evaluation","author":"Wang","year":"2023","journal-title":"arXiv:2306.09968"},{"key":"ref117","article-title":"IvyGPT: InteractiVe Chinese pathwaY language model in medical domain","author":"Wang","year":"2023","journal-title":"arXiv:2307.10512"},{"key":"ref118","article-title":"XrayGLM","author":"Wang"},{"key":"ref119","doi-asserted-by":"publisher","DOI":"10.1016\/j.cbpa.2004.06.001"},{"key":"ref120","doi-asserted-by":"publisher","DOI":"10.1145\/3293318"},{"key":"ref121","doi-asserted-by":"publisher","DOI":"10.21437\/interspeech.2017-1452"},{"key":"ref122","doi-asserted-by":"publisher","DOI":"10.1126\/science.1411538"},{"key":"ref123","doi-asserted-by":"publisher","DOI":"10.1002\/cpz1.273"},{"key":"ref124","article-title":"DoctorGLM: Fine-tuning your Chinese doctor is not a herculean task","author":"Xiong","year":"2023","journal-title":"arXiv:2304.01097"},{"key":"ref125","first-page":"38749","article-title":"ProtST: Multi-modality learning of protein sequences and biomedical texts","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Xu"},{"key":"ref126","doi-asserted-by":"publisher","DOI":"10.1111\/cts.12980"},{"key":"ref127","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-022-00534-z"},{"key":"ref128","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2017.2787158"},{"key":"ref129","article-title":"Interactive molecular discovery with natural language","author":"Zeng","year":"2023","journal-title":"arXiv:2306.11976"},{"key":"ref130","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.725"},{"key":"ref131","article-title":"BiomedGPT: A generalist vision-language foundation model for diverse biomedical tasks","author":"Zhang","year":"2023","journal-title":"arXiv:2305.17100"},{"key":"ref132","doi-asserted-by":"publisher","DOI":"10.1515\/jib-2019-0027"},{"key":"ref133","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbab152"},{"key":"ref134","doi-asserted-by":"publisher","DOI":"10.1093\/nar\/gkad1031"},{"key":"ref135","doi-asserted-by":"publisher","DOI":"10.3389\/fphar.2023.1194216"},{"key":"ref136","doi-asserted-by":"publisher","DOI":"10.18293\/JVLC2020-N2-010"},{"key":"ref137","doi-asserted-by":"publisher","DOI":"10.1007\/s41095-023-0364-2"},{"key":"ref138","doi-asserted-by":"publisher","DOI":"10.1177\/10943420231201154"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/10930479.pdf?arnumber=10930479","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T23:42:59Z","timestamp":1743464579000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10930479\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":138,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3552256","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}