{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T04:02:14Z","timestamp":1770350534084,"version":"3.49.0"},"reference-count":77,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Satellite Network of Experts (SatNEx) V Project funded by European Space Agency","award":["4000120693\/17\/NL\/CLP"],"award-info":[{"award-number":["4000120693\/17\/NL\/CLP"]}]},{"name":"Transversal Skills in Applied Artificial Intelligence (TSAAI) co-funded by European Commission","award":["2021-1-ES01-KA220-HED-000030125"],"award-info":[{"award-number":["2021-1-ES01-KA220-HED-000030125"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3605022","type":"journal-article","created":{"date-parts":[[2025,9,2]],"date-time":"2025-09-02T17:31:52Z","timestamp":1756834312000},"page":"155675-155696","source":"Crossref","is-referenced-by-count":1,"title":["Adapting LLMs for Satellite Communications: Methodology, Challenges, and Impact"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8749-8934","authenticated-orcid":false,"given":"Alejandro","family":"Mozo","sequence":"first","affiliation":[{"name":"Department of Languages and Computer Sciences, University of Malaga (Andalucia Tech), Malaga, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4743-2807","authenticated-orcid":false,"given":"Sergio","family":"G\u00e1lvez","sequence":"additional","affiliation":[{"name":"Department of Languages and Computer Sciences, University of Malaga (Andalucia Tech), Malaga, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0206-6973","authenticated-orcid":false,"given":"Ioannis T.","family":"Christou","sequence":"additional","affiliation":[{"name":"Department of Information Technology, Cybersecurity and Computer Science, The American College of Greece, Aghia Paraskevi, Greece"}]},{"given":"Dimitrios","family":"Vogiatzis","sequence":"additional","affiliation":[{"name":"Department of Information Technology, Cybersecurity and Computer Science, The American College of Greece, Aghia Paraskevi, Greece"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-0056-0238","authenticated-orcid":false,"given":"Tom\u00e1s","family":"Navarro","sequence":"additional","affiliation":[{"name":"European Centre for Space Applications and Telecommunications (ECSAT), European Space Agency, Didcot, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0778-6838","authenticated-orcid":false,"given":"Francisco L.","family":"Valverde","sequence":"additional","affiliation":[{"name":"Department of Languages and Computer Sciences, University of Malaga (Andalucia Tech), Malaga, Spain"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1039\/d3dd00112a"},{"key":"ref2","article-title":"Fine-tuning or fine-failing? Debunking performance myths in large language models","author":"Barnett","year":"2024","journal-title":"arXiv:2406.11201"},{"key":"ref3","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Brown"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.634"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1186\/s40163-025-00248-8"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1145\/3594536.3595165"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/3764579"},{"key":"ref9","article-title":"Longformer: The long-document transformer","author":"Beltagy","year":"2020","journal-title":"arXiv:2004.05150"},{"key":"ref10","article-title":"Publicly available clinical BERT embeddings","author":"Alsentzer","year":"2019","journal-title":"arXiv:1904.03323"},{"key":"ref11","article-title":"Finetuned language models are zero-shot learners","author":"Wei","year":"2021","journal-title":"arXiv:2109.01652"},{"key":"ref12","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv:1810.04805"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1186\/s12864-021-07490-9"},{"key":"ref14","article-title":"Investigating the use of federated learning to enable collaborative AI model training while preserving data privacy in industrial IoTT settings","author":"Luz","year":"2024"},{"key":"ref15","article-title":"Fine-tuning large language models for domain adaptation: Exploration of training strategies, scaling, model merging and synergistic capabilities","author":"Lu","year":"2024","journal-title":"arXiv:2409.03444"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.7759\/cureus.57728"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijinfomgt.2019.08.002"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.23919\/icn.2021.0015"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/AERO47225.2020.9172454"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0211813"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/access.2021.3064976"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.epsr.2023.110068"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ojcoms.2024.3522103"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1016\/j.adhoc.2024.103645"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/iiswc63097.2024.00027"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/w18-5406"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1907.11692"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-04346-8_62"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/s10032-015-0249-8"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1186\/1751-0473-7-7"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.5220\/0010524503590366"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.1c01198"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/JCDL.2017.7991564"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-28032-0_31"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/icdar.2019.00029"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1017\/s1351324923000049"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-76941-7_13"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-40501-3_15"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.577"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/sp46215.2023.10179300"},{"key":"ref41","volume-title":"The Empirical Impact of Data Sanitization on Language Models","author":"Pal","year":"2024"},{"key":"ref42","article-title":"Fine-tuning and utilization methods of domain-specific LLMs","author":"Jeong","year":"2024","journal-title":"arXiv:2401.02981"},{"key":"ref43","article-title":"Galactica: A large language model for science","author":"Taylor","year":"2022","journal-title":"arXiv:2211.09085"},{"key":"ref44","article-title":"LoRA: Low-rank adaptation of large language models","author":"Hu","year":"2021","journal-title":"arXiv:2106.09685"},{"key":"ref45","first-page":"10088","article-title":"QLoRA: Efficient finetuning of quantized LLMs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Dettmers"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/sc41404.2022.00051"},{"key":"ref47","first-page":"31094","article-title":"FlexGen: high-throughput generative inference of large language models with a single GPU","volume-title":"Proc. Mach. Learn. Res.","author":"Sheng"},{"key":"ref48","first-page":"489","article-title":"PetS: A unified framework for parameter-efficient transformers serving","volume-title":"Proc. USENIX Annu. Tech. Conf.","author":"Zhou"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-industry.15"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/cloud62652.2024.00028"},{"key":"ref51","article-title":"70B-parameter large language models in Japanese medical question-answering","author":"Sukeda","year":"2024","journal-title":"arXiv:2406.14882"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"key":"ref53","first-page":"74","article-title":"ROUGE: A package for automatic evaluation of summaries","volume-title":"Proc. Workshop Text Summarization Branches Out","author":"Lin"},{"key":"ref54","first-page":"65","article-title":"METEOR: An automatic metric for MT evaluation with high levels of correlation with human judgments","volume-title":"Proc. ACL Workshop Intrinsic Extrinsic Eval. Measures Mach. Transl. Summarization","author":"Banerjee"},{"key":"ref55","first-page":"1","article-title":"BERTScore: Evaluating text generation with BERT","volume-title":"Proc. 8th Int. Conf. Learn. Represent.","author":"Zhang"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.704"},{"key":"ref57","article-title":"Judging LLM-as-a-judge with MT-bench and chatbot arena","author":"Zheng","year":"2023","journal-title":"arXiv:2306.05685"},{"key":"ref58","first-page":"1","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Lee"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-eacl.83"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2025.findings-acl.306"},{"key":"ref61","article-title":"GPTScore: Evaluate as you desire","author":"Fu","year":"2023","journal-title":"arXiv:2302.04166"},{"key":"ref62","article-title":"Prometheus: Inducing fine-grained evaluation capability in language models","author":"Kim","year":"2023","journal-title":"arXiv:2310.08491"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.891"},{"key":"ref64","article-title":"PandaLM: An automatic evaluation benchmark for LLM instruction tuning optimization","author":"Wang","year":"2023","journal-title":"arXiv:2306.05087"},{"key":"ref65","first-page":"1","article-title":"ChatEval: Towards better LLM-based evaluators through multi-agent debate","volume-title":"Proc. 12th Int. Conf. Learn. Represent.","author":"Chan"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.741"},{"key":"ref67","article-title":"OpenFactCheck: Building, benchmarking customized fact-checking systems and evaluating the factuality of claims and LLMs","author":"Wang","year":"2024","journal-title":"arXiv:2405.05583"},{"key":"ref68","article-title":"Evaluating the factuality of large language models using large-scale knowledge graphs","author":"Liu","year":"2024","journal-title":"arXiv:2404.00942"},{"key":"ref69","article-title":"LLMs-as-judges: A comprehensive survey on LLM-based evaluation methods","author":"Li","year":"2024","journal-title":"arXiv:2412.05579"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/mwc.2018.1800155"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1145\/3231884.3231901"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1109\/lwc.2021.3083267"},{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1016\/j.asej.2019.08.006"},{"key":"ref74","article-title":"Artificial intelligence for satellite communication and non-terrestrial networks: A survey","author":"Fontanesi","year":"2023","journal-title":"arXiv:2304.13008"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/comst.2020.3028247"},{"key":"ref76","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Ouyang"},{"key":"ref77","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330955"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/11146720.pdf?arnumber=11146720","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T17:48:00Z","timestamp":1759859280000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11146720\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":77,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3605022","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}