{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T14:24:41Z","timestamp":1773930281810,"version":"3.50.1"},"reference-count":64,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","award":["RS-2023-00251105"],"award-info":[{"award-number":["RS-2023-00251105"]}],"id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100014188","name":"Korea Ministry of Science and ICT","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100014188","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004085","name":"Ministry of Education, Science and Technology","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100004085","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010418","name":"Institute for Information and Communications Technology Promotion","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100010418","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100019403","name":"Information Technology Research Centre","doi-asserted-by":"publisher","award":["IITP-RS-2025-00436248"],"award-info":[{"award-number":["IITP-RS-2025-00436248"]}],"id":[{"id":"10.13039\/501100019403","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Engineering Applications of Artificial Intelligence"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1016\/j.engappai.2025.113548","type":"journal-article","created":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T07:14:14Z","timestamp":1765782854000},"page":"113548","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"PB","title":["Multiple quality-of-services optimization in space\u2013air\u2013ground integrated network: Centralized and decentralized deep reinforcement learning approaches"],"prefix":"10.1016","volume":"165","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9049-4459","authenticated-orcid":false,"given":"Sengly","family":"Muy","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0006-5221-5945","authenticated-orcid":false,"given":"Vitou","family":"That","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7372-5692","authenticated-orcid":false,"given":"Jung-Ryun","family":"Lee","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.engappai.2025.113548_b1","doi-asserted-by":"crossref","DOI":"10.1109\/JIOT.2024.3370192","article-title":"Long-term throughput maximization in wireless powered communication networks: A multi-task DRL approach","author":"Ahmadian","year":"2024","journal-title":"IEEE Internet Things J."},{"key":"10.1016\/j.engappai.2025.113548_b2","doi-asserted-by":"crossref","DOI":"10.1109\/ACCESS.2024.3525263","article-title":"Machine learning applications in energy harvesting internet of things networks: A review","author":"Alamu","year":"2025","journal-title":"IEEE Access"},{"key":"10.1016\/j.engappai.2025.113548_b3","series-title":"International Conference on Machine Learning","first-page":"243","article-title":"An alternative softmax operator for reinforcement learning","author":"Asadi","year":"2017"},{"issue":"5","key":"10.1016\/j.engappai.2025.113548_b4","doi-asserted-by":"crossref","first-page":"483","DOI":"10.1002\/wcm.72","article-title":"A survey of mobility models for ad hoc network research","volume":"2","author":"Camp","year":"2002","journal-title":"Wirel. Commun. Mob. Comput."},{"key":"10.1016\/j.engappai.2025.113548_b5","series-title":"Deep deterministic policy gradient for urban traffic light control","author":"Casas","year":"2017"},{"key":"10.1016\/j.engappai.2025.113548_b6","article-title":"Boltzmann exploration done right","volume":"30","author":"Cesa-Bianchi","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"2","key":"10.1016\/j.engappai.2025.113548_b7","doi-asserted-by":"crossref","first-page":"934","DOI":"10.1109\/TGCN.2021.3061631","article-title":"Reinforcement learning-based energy-efficient data access for airborne users in civil aircrafts-enabled SAGIN","volume":"5","author":"Chen","year":"2021","journal-title":"IEEE Trans. Green Commun. Netw."},{"issue":"8","key":"10.1016\/j.engappai.2025.113548_b8","doi-asserted-by":"crossref","first-page":"5368","DOI":"10.1109\/TCOMM.2022.3186997","article-title":"Robust task scheduling for delay-aware IoT applications in civil aircraft-augmented SAGIN","volume":"70","author":"Chen","year":"2022","journal-title":"IEEE Trans. Commun."},{"issue":"8","key":"10.1016\/j.engappai.2025.113548_b9","doi-asserted-by":"crossref","first-page":"4988","DOI":"10.1109\/TCOMM.2023.3280219","article-title":"Coverage analysis of sagin with sectorized beam pattern under shadowed-rician fading channels","volume":"71","author":"Chen","year":"2023","journal-title":"IEEE Trans. Commun."},{"key":"10.1016\/j.engappai.2025.113548_b10","series-title":"International Conference on Machine Learning","first-page":"4666","article-title":"Guarantees for epsilon-greedy reinforcement learning with function approximation","author":"Dann","year":"2022"},{"key":"10.1016\/j.engappai.2025.113548_b11","doi-asserted-by":"crossref","first-page":"165","DOI":"10.1023\/A:1019181312967","article-title":"Characterization of user mobility in low earth orbit mobile satellite systems","volume":"6","author":"Del Re","year":"2000","journal-title":"Wirel. Netw."},{"issue":"3","key":"10.1016\/j.engappai.2025.113548_b12","doi-asserted-by":"crossref","first-page":"1802","DOI":"10.1109\/TNSE.2022.3153480","article-title":"Network selection based on evolutionary game and deep reinforcement learning in space-air-ground integrated network","volume":"9","author":"Fan","year":"2022","journal-title":"IEEE Trans. Netw. Sci. Eng."},{"issue":"9","key":"10.1016\/j.engappai.2025.113548_b13","doi-asserted-by":"crossref","first-page":"9265","DOI":"10.1109\/TVT.2020.2972133","article-title":"Joint 3D trajectory design and time allocation for UAV-enabled wireless power transfer networks","volume":"69","author":"Feng","year":"2020","journal-title":"IEEE Trans. Veh. Technol."},{"issue":"2","key":"10.1016\/j.engappai.2025.113548_b14","doi-asserted-by":"crossref","first-page":"1226","DOI":"10.1109\/COMST.2021.3063822","article-title":"Single and multi-agent deep reinforcement learning for AI-enabled wireless networks: A tutorial","volume":"23","author":"Feriani","year":"2021","journal-title":"IEEE Commun. Surv. & Tutorials"},{"key":"10.1016\/j.engappai.2025.113548_b15","article-title":"Learning to communicate with deep multi-agent reinforcement learning","volume":"29","author":"Foerster","year":"2016","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"11","key":"10.1016\/j.engappai.2025.113548_b16","doi-asserted-by":"crossref","first-page":"1989","DOI":"10.1109\/LCOMM.2019.2936393","article-title":"Deep learning based channel estimation for massive MIMO with mixed-resolution ADCs","volume":"23","author":"Gao","year":"2019","journal-title":"IEEE Commun. Lett."},{"issue":"5","key":"10.1016\/j.engappai.2025.113548_b17","doi-asserted-by":"crossref","first-page":"2816","DOI":"10.1109\/TCOMM.2023.3251342","article-title":"Jointly optimized beamforming and power allocation for full-duplex cell-free NOMA in space-ground integrated networks","volume":"71","author":"Gao","year":"2023","journal-title":"IEEE Trans. Commun."},{"key":"10.1016\/j.engappai.2025.113548_b18","doi-asserted-by":"crossref","first-page":"129358","DOI":"10.1109\/ACCESS.2021.3113501","article-title":"Deep reinforcement learning for energy-efficient multi-channel transmissions in 5G cognitive hetnets: Centralized, decentralized and transfer learning based solutions","volume":"9","author":"Giannopoulos","year":"2021","journal-title":"IEEE Access"},{"key":"10.1016\/j.engappai.2025.113548_b19","series-title":"International Conference on Machine Learning","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","author":"Haarnoja","year":"2018"},{"key":"10.1016\/j.engappai.2025.113548_b20","doi-asserted-by":"crossref","first-page":"17","DOI":"10.1016\/j.ins.2022.10.022","article-title":"Entropy regularized actor-critic based multi-agent deep reinforcement learning for stochastic games","volume":"617","author":"Hao","year":"2022","journal-title":"Inform. Sci."},{"issue":"1","key":"10.1016\/j.engappai.2025.113548_b21","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1109\/TMC.2022.3222848","article-title":"Balancing total energy consumption and mean makespan in data offloading for space-air-ground integrated networks","volume":"23","author":"He","year":"2022","journal-title":"IEEE Trans. Mob. Comput."},{"key":"10.1016\/j.engappai.2025.113548_b22","article-title":"Deep q-learning from demonstrations","volume":"vol. 32","author":"Hester","year":"2018"},{"key":"10.1016\/j.engappai.2025.113548_b23","series-title":"Machine Learning: Master Supervised and Unsupervised Learning Algorithms with Real Examples (English Edition)","author":"Hiran","year":"2021"},{"key":"10.1016\/j.engappai.2025.113548_b24","series-title":"Action noise in off-policy deep reinforcement learning: Impact on exploration and performance","author":"Hollenstein","year":"2022"},{"key":"10.1016\/j.engappai.2025.113548_b25","series-title":"Scalable multi-agent reinforcement learning algorithm for wireless networks","author":"Hu","year":"2021"},{"issue":"6","key":"10.1016\/j.engappai.2025.113548_b26","doi-asserted-by":"crossref","first-page":"451","DOI":"10.1002\/sat.1250","article-title":"Velocity-aware handover prediction in LEO satellite communication networks","volume":"36","author":"Hu","year":"2018","journal-title":"Int. J. Satell. Commun. Netw."},{"issue":"1","key":"10.1016\/j.engappai.2025.113548_b27","article-title":"Deep learning for dynamic resource management in 5G networks: A review","volume":"14","author":"Hussein","year":"2025","journal-title":"Indones. J. Comput. Sci."},{"issue":"13","key":"10.1016\/j.engappai.2025.113548_b28","doi-asserted-by":"crossref","first-page":"10516","DOI":"10.1109\/JIOT.2021.3121760","article-title":"Toward data collection and transmission in 6G space\u2013air\u2013ground integrated networks: Cooperative HAP and LEO satellite schemes","volume":"9","author":"Jia","year":"2021","journal-title":"IEEE Internet Things J."},{"key":"10.1016\/j.engappai.2025.113548_b29","series-title":"2023 IEEE International Conference on Paradigm Shift in Information Technologies with Innovative Applications in Global Scenario","first-page":"392","article-title":"Novel deep learning network architectures for time series forecasting","author":"Katariya","year":"2023"},{"key":"10.1016\/j.engappai.2025.113548_b30","series-title":"Learning to schedule communication in multi-agent reinforcement learning","author":"Kim","year":"2019"},{"issue":"1","key":"10.1016\/j.engappai.2025.113548_b31","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/MAP.2016.2630035","article-title":"Path-loss modeling for wireless sensor networks: A review of models and comparative evaluations","volume":"59","author":"Kurt","year":"2017","journal-title":"IEEE Antennas Propag. Mag."},{"key":"10.1016\/j.engappai.2025.113548_b32","article-title":"Scaling up multi-agent reinforcement learning: An extensive survey on scalability issues","author":"Liu","year":"2024","journal-title":"IEEE Access"},{"issue":"9","key":"10.1016\/j.engappai.2025.113548_b33","doi-asserted-by":"crossref","first-page":"9631","DOI":"10.1109\/TII.2022.3227652","article-title":"A DRL-based decentralized computation offloading method: An example of an intelligent manufacturing scenario","volume":"19","author":"Lu","year":"2022","journal-title":"IEEE Trans. Ind. Informatics"},{"issue":"9","key":"10.1016\/j.engappai.2025.113548_b34","first-page":"1006","article-title":"Efficient and scalable reinforcement learning for large-scale network control","volume":"6","author":"Ma","year":"2024","journal-title":"Nat. Mach. Intell."},{"key":"10.1016\/j.engappai.2025.113548_b35","series-title":"Satellite Communications Systems: Systems, Techniques and Technology","author":"Maral","year":"2020"},{"key":"10.1016\/j.engappai.2025.113548_b36","series-title":"Grokking Deep Reinforcement Learning","author":"Morales","year":"2020"},{"issue":"2","key":"10.1016\/j.engappai.2025.113548_b37","doi-asserted-by":"crossref","first-page":"142","DOI":"10.23919\/ICN.2020.0009","article-title":"Space-air-ground integrated vehicular network for connected and automated vehicles: Challenges and solutions","volume":"1","author":"Niu","year":"2020","journal-title":"Intell. Converg. Networks"},{"key":"10.1016\/j.engappai.2025.113548_b38","doi-asserted-by":"crossref","DOI":"10.3389\/fsufs.2024.1450653","article-title":"Enhancement of the energy and exergy analysis capabilities of the yoghurt process: a case study of the dairy industry","volume":"8","author":"Oztuna Taner","year":"2024","journal-title":"Front. Sustain. Food Syst."},{"key":"10.1016\/j.engappai.2025.113548_b39","doi-asserted-by":"crossref","DOI":"10.3389\/fsufs.2024.1344370","article-title":"Dairy factory milk product processing and sustainable of the shelf-life extension with artificial intelligence: a model study","volume":"8","author":"Oztuna Taner","year":"2024","journal-title":"Front. Sustain. Food Syst."},{"issue":"5","key":"10.1016\/j.engappai.2025.113548_b40","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3453160","article-title":"Hierarchical reinforcement learning: A comprehensive survey","volume":"54","author":"Pateria","year":"2021","journal-title":"ACM Comput. Surv."},{"key":"10.1016\/j.engappai.2025.113548_b41","series-title":"2012 IEEE International Symposium on Information Theory Proceedings","first-page":"2446","article-title":"Applications of the shannon-hartley theorem to data streams and sparse recovery","author":"Price","year":"2012"},{"key":"10.1016\/j.engappai.2025.113548_b42","series-title":"2023 International Conference on Power Energy, Environment & Intelligent Control","first-page":"1168","article-title":"An implementation of multi based and objective based energy management (Saving) system","author":"Pushparajesh","year":"2023"},{"key":"10.1016\/j.engappai.2025.113548_b43","series-title":"2024 3rd International Conference for Advancement in Technology","first-page":"1","article-title":"Satellite communications performance analysis in the presence of atmospheric attenuation and rain fading","author":"Pushparajesh","year":"2024"},{"key":"10.1016\/j.engappai.2025.113548_b44","series-title":"Gaussian process policy optimization","author":"Rao","year":"2020"},{"issue":"178","key":"10.1016\/j.engappai.2025.113548_b45","first-page":"1","article-title":"Monotonic value function factorisation for deep multi-agent reinforcement learning","volume":"21","author":"Rashid","year":"2020","journal-title":"J. Mach. Learn. Res."},{"issue":"3","key":"10.1016\/j.engappai.2025.113548_b46","doi-asserted-by":"crossref","first-page":"2830","DOI":"10.1109\/TITS.2021.3119921","article-title":"Speech emotion recognition enhanced traffic efficiency solution for autonomous vehicles in a 5G-enabled space\u2013air\u2013ground integrated intelligent transportation system","volume":"23","author":"Tan","year":"2021","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.engappai.2025.113548_b47","doi-asserted-by":"crossref","first-page":"261","DOI":"10.1016\/j.fbp.2024.11.028","article-title":"Improve the energy efficiency of the fruit freeze-drying through the predictive analysis","volume":"149","author":"Taner","year":"2025","journal-title":"Food Bioprod. Process."},{"key":"10.1016\/j.engappai.2025.113548_b48","series-title":"Advanced Materials-Based Thermally Enhanced Phase Change Materials","first-page":"269","article-title":"Application of artificial intelligence techniques for heat exchanger predictions in food industry","author":"Taner","year":"2024"},{"key":"10.1016\/j.engappai.2025.113548_b49","series-title":"Deep reinforcement learning for robotics: A survey of real-world successes","author":"Tang","year":"2024"},{"key":"10.1016\/j.engappai.2025.113548_b50","series-title":"Information Systems Architecture and Technology: Proceedings of 39th International Conference on Information Systems Architecture and Technology\u2013ISAT 2018: Part II","first-page":"173","article-title":"Energy consumption in unmanned aerial vehicles: A review of energy consumption models and their relation to the UAV routing","author":"Thibbotuwawa","year":"2019"},{"key":"10.1016\/j.engappai.2025.113548_b51","series-title":"5G and Satellite RF and Optical Integration","author":"Varrall","year":"2022"},{"issue":"3","key":"10.1016\/j.engappai.2025.113548_b52","doi-asserted-by":"crossref","first-page":"261","DOI":"10.1038\/s41592-019-0686-2","article-title":"SciPy 1.0: fundamental algorithms for scientific computing in python","volume":"17","author":"Virtanen","year":"2020","journal-title":"Nature Methods"},{"key":"10.1016\/j.engappai.2025.113548_b53","article-title":"Scheduling of real-time wireless flows: A comparative study of centralized and decentralized reinforcement learning approaches","author":"Wang","year":"2024","journal-title":"IEEE\/ACM Trans. Netw."},{"issue":"3","key":"10.1016\/j.engappai.2025.113548_b54","doi-asserted-by":"crossref","first-page":"434","DOI":"10.1109\/JSTSP.2021.3136027","article-title":"Incorporating distributed DRL into storage resource optimization of space-air-ground integrated wireless communication network","volume":"16","author":"Wang","year":"2021","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"10.1016\/j.engappai.2025.113548_b55","series-title":"2021 IEEE\/CIC International Conference on Communications in China","first-page":"315","article-title":"Survey on Space-air-ground Integrated Networks in 6G","author":"Wang","year":"2021"},{"key":"10.1016\/j.engappai.2025.113548_b56","series-title":"2024 IEEE Wireless Communications and Networking Conference","first-page":"1","article-title":"Deep reinforcement learning-based adaptive offloading algorithm for wireless power transfer-aided mobile edge computing","author":"Wu","year":"2024"},{"key":"10.1016\/j.engappai.2025.113548_b57","series-title":"IEEE International Conference on Communications, 2003, Vol. 5","first-page":"3524","article-title":"Statistical simulation models for Rayleigh and rician fading","author":"Xiao","year":"2003"},{"key":"10.1016\/j.engappai.2025.113548_b58","first-page":"24611","article-title":"The surprising effectiveness of ppo in cooperative multi-agent games","volume":"35","author":"Yu","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.engappai.2025.113548_b59","series-title":"2018 IEEE\/ACM 26th International Symposium on Quality of Service","first-page":"1","article-title":"Improved adam optimizer for deep neural networks","author":"Zhang","year":"2018"},{"issue":"4","key":"10.1016\/j.engappai.2025.113548_b60","doi-asserted-by":"crossref","first-page":"5687","DOI":"10.1109\/TPWRS.2023.3336614","article-title":"DNN assisted projection based deep reinforcement learning for safe control of distribution grids","volume":"39","author":"Zhang","year":"2023","journal-title":"IEEE Trans. Power Syst."},{"issue":"3","key":"10.1016\/j.engappai.2025.113548_b61","doi-asserted-by":"crossref","first-page":"3348","DOI":"10.1109\/TNSM.2022.3232414","article-title":"Distributed deep reinforcement learning assisted resource allocation algorithm for space-air-ground integrated networks","volume":"20","author":"Zhang","year":"2022","journal-title":"IEEE Trans. Netw. Serv. Manag."},{"issue":"23","key":"10.1016\/j.engappai.2025.113548_b62","doi-asserted-by":"crossref","first-page":"20472","DOI":"10.1109\/JIOT.2023.3287737","article-title":"Multiagent reinforcement learning-based orbital edge offloading in SAGIN supporting internet of remote things","volume":"10","author":"Zhang","year":"2023","journal-title":"IEEE Internet Things J."},{"key":"10.1016\/j.engappai.2025.113548_b63","doi-asserted-by":"crossref","unstructured":"Zhang, J., Zheng, Y., Qi, D., Li, R., Yi, X., 2016. DNN-based prediction model for spatio-temporal data. In: Proceedings of the 24th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems. pp. 1\u20134.","DOI":"10.1145\/2996913.2997016"},{"issue":"1","key":"10.1016\/j.engappai.2025.113548_b64","doi-asserted-by":"crossref","first-page":"4","DOI":"10.1007\/s10458-023-09633-6","article-title":"A survey of multi-agent deep reinforcement learning with communication","volume":"38","author":"Zhu","year":"2024","journal-title":"Auton. Agents Multi-Agent Syst."}],"container-title":["Engineering Applications of Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625035791?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625035791?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T11:28:33Z","timestamp":1773919713000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0952197625035791"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":64,"alternative-id":["S0952197625035791"],"URL":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113548","relation":{},"ISSN":["0952-1976"],"issn-type":[{"value":"0952-1976","type":"print"}],"subject":[],"published":{"date-parts":[[2026,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Multiple quality-of-services optimization in space\u2013air\u2013ground integrated network: Centralized and decentralized deep reinforcement learning approaches","name":"articletitle","label":"Article Title"},{"value":"Engineering Applications of Artificial Intelligence","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113548","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"113548"}}