{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T22:49:43Z","timestamp":1726094983445},"reference-count":48,"publisher":"IGI Global","issue":"1","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,1]]},"abstract":"<jats:p>Heterogeneous networks (HetNets) can equalize traffic loads and cut down the cost of deploying cells. Thus, it is regarded to be the significant technique of the next-generation communication networks. Due to the non-convexity nature of the channel allocation problem in HetNets, it is difficult to design an optimal approach for allocating channels. To ensure the user quality of service as well as the long-term total network utility, this article proposes a new method through utilizing multi-agent reinforcement learning. Moreover, for the purpose of solving computational complexity problem caused by the large action space, deep reinforcement learning is put forward to learn optimal policy. A nearly-optimal solution with high efficiency and rapid convergence speed could be obtained by this learning method. Simulation results reveal that this new method has the best performance than other methods.<\/jats:p>","DOI":"10.4018\/ijmcmc.2020010102","type":"journal-article","created":{"date-parts":[[2020,2,7]],"date-time":"2020-02-07T19:25:21Z","timestamp":1581103521000},"page":"23-41","source":"Crossref","is-referenced-by-count":4,"title":["Multi-Agent Actor Critic for Channel Allocation in Heterogeneous Networks"],"prefix":"10.4018","volume":"11","author":[{"given":"Nan","family":"Zhao","sequence":"first","affiliation":[{"name":"Hubei University of Technology, China"}]},{"given":"Zehua","family":"Liu","sequence":"additional","affiliation":[{"name":"Hubei University of Technology, China"}]},{"given":"Yiqiang","family":"Cheng","sequence":"additional","affiliation":[{"name":"Hubei University of Technology, China"}]},{"given":"Chao","family":"Tian","sequence":"additional","affiliation":[{"name":"Hubei University of Technology, China"}]}],"member":"2432","reference":[{"key":"IJMCMC.2020010102-0","doi-asserted-by":"publisher","DOI":"10.1049\/iet-com.2019.0383"},{"key":"IJMCMC.2020010102-1","doi-asserted-by":"publisher","DOI":"10.1504\/IJSSC.2015.070953"},{"key":"IJMCMC.2020010102-2","doi-asserted-by":"publisher","DOI":"10.1109\/TCOMM.2016.2593468"},{"key":"IJMCMC.2020010102-3","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2939827"},{"key":"IJMCMC.2020010102-4","doi-asserted-by":"publisher","DOI":"10.1504\/IJSSC.2017.089003"},{"key":"IJMCMC.2020010102-5","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2019.2900035"},{"key":"IJMCMC.2020010102-6","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2019.2933762"},{"key":"IJMCMC.2020010102-7","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2013.2268923"},{"key":"IJMCMC.2020010102-8","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2878752"},{"key":"IJMCMC.2020010102-9","doi-asserted-by":"publisher","DOI":"10.1145\/1143844.1143877"},{"key":"IJMCMC.2020010102-10","doi-asserted-by":"crossref","unstructured":"D\u2019Eramo, C., Nuara, A., Pirotta, M., & Restelli, M. (2017). Estimating the maximum expected value in continuous reinforcement learning problems. In Proceedings of the31st AAAI Conference on Artificial Intelligence (pp. 1840-1846). AAAI Press.","DOI":"10.1609\/aaai.v31i1.10887"},{"key":"IJMCMC.2020010102-11","unstructured":"Dulac-Arnold, G., Evans, R., van Hasselt, H., Sunehag, P., Lillicrap, T., Hunt, J., . . . Coppin, B. (2016). Deep reinforcement learning in large discrete action spaces. In Proceedings of theInternational Conference on Machine Learning (ICML). Academic Press."},{"key":"IJMCMC.2020010102-12","first-page":"7","article-title":"Mixed reinforcement learning for partially observable Markov decision process.","author":"L.Dung","year":"2006","journal-title":"Proceedings of the 2007 IEEE International Symposium on Computational Intelligence in Robotics and Automation (CIRA)"},{"key":"IJMCMC.2020010102-13","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2015.2416990"},{"key":"IJMCMC.2020010102-14","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2013.2255286"},{"key":"IJMCMC.2020010102-15","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-68321-8_8"},{"key":"IJMCMC.2020010102-16","doi-asserted-by":"publisher","DOI":"10.1109\/JCN.2018.000073"},{"key":"IJMCMC.2020010102-17","doi-asserted-by":"publisher","DOI":"10.1007\/s41650-017-0002-1"},{"key":"IJMCMC.2020010102-18","doi-asserted-by":"crossref","first-page":"354","DOI":"10.1007\/978-3-319-41649-6_36","article-title":"Ideas for a reinforcement learning algorithm that learns programs.","author":"S.Katayama","year":"2016","journal-title":"Proceedings of the International Conference on Artificial General Intelligence"},{"key":"IJMCMC.2020010102-19","doi-asserted-by":"publisher","DOI":"10.4018\/IJMCMC.2019040101"},{"key":"IJMCMC.2020010102-20","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2017.2744620"},{"key":"IJMCMC.2020010102-21","doi-asserted-by":"publisher","DOI":"10.1504\/IJSSC.2017.086811"},{"issue":"12","key":"IJMCMC.2020010102-22","first-page":"3136","article-title":"Shallow updates for deep reinforcement learning.","author":"N.Levine","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"IJMCMC.2020010102-23","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"IJMCMC.2020010102-24","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2930115"},{"key":"IJMCMC.2020010102-25","doi-asserted-by":"publisher","DOI":"10.1109\/TGCN.2018.2844301"},{"key":"IJMCMC.2020010102-26","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2018.2874229"},{"key":"IJMCMC.2020010102-27","doi-asserted-by":"publisher","DOI":"10.1109\/TII.2018.2791424"},{"key":"IJMCMC.2020010102-28","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2019.2924579"},{"key":"IJMCMC.2020010102-29","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2015.12.013"},{"key":"IJMCMC.2020010102-30","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553504"},{"key":"IJMCMC.2020010102-31","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2935545"},{"key":"IJMCMC.2020010102-32","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2016.2616384"},{"key":"IJMCMC.2020010102-33","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992698"},{"key":"IJMCMC.2020010102-34","doi-asserted-by":"publisher","DOI":"10.1109\/JCN.2018.000019"},{"key":"IJMCMC.2020010102-35","doi-asserted-by":"publisher","DOI":"10.1109\/LCOMM.2018.2875689"},{"key":"IJMCMC.2020010102-36","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2017.2769644"},{"key":"IJMCMC.2020010102-37","doi-asserted-by":"publisher","DOI":"10.1109\/TCOMM.2018.2831207"},{"key":"IJMCMC.2020010102-38","doi-asserted-by":"publisher","DOI":"10.1109\/CC.2018.8438283"},{"key":"IJMCMC.2020010102-39","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2018.2873139"},{"key":"IJMCMC.2020010102-40","doi-asserted-by":"publisher","DOI":"10.4018\/IJMCMC.2019010103"},{"key":"IJMCMC.2020010102-41","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2017.2786255"},{"key":"IJMCMC.2020010102-42","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2018.2865817"},{"key":"IJMCMC.2020010102-43","doi-asserted-by":"publisher","DOI":"10.1109\/GLOCOM.2017.8254092"},{"key":"IJMCMC.2020010102-44","doi-asserted-by":"crossref","unstructured":"Zhao, N., Fan, P. F., He, X., Fan, M. L., & Tian, C. (2019a). Two-stage dynamic contract deign under asymmetric information in cooperative communication. In Proceedings of the 14th International Conference on Broad-Band Wireless Computing, Communication and Applications (BWCCA-2019). Academic Press.","DOI":"10.1007\/978-3-030-33506-9_57"},{"key":"IJMCMC.2020010102-45","doi-asserted-by":"crossref","unstructured":"Zhao, N., Fan, M. L., Tian, C., Fan, P. F., & He, X. (2019b). Multi-dimensional contract incentive design for mobile crowdsourcing networks. In Proceedings of the 14th International Conference on Broad-Band Wireless Computing, Communication and Applications (BWCCA-2019). Academic Press.","DOI":"10.1007\/978-3-030-33506-9_51"},{"key":"IJMCMC.2020010102-46","doi-asserted-by":"publisher","DOI":"10.4018\/IJMCMC.2018100103"},{"key":"IJMCMC.2020010102-47","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2018.2864415"}],"container-title":["International Journal of Mobile Computing and Multimedia Communications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.igi-global.com\/viewtitle.aspx?TitleId=248450","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,14]],"date-time":"2022-10-14T22:55:43Z","timestamp":1665788143000},"score":1,"resource":{"primary":{"URL":"http:\/\/services.igi-global.com\/resolvedoi\/resolve.aspx?doi=10.4018\/IJMCMC.2020010102"}},"subtitle":[""],"short-title":[],"issued":{"date-parts":[[2020,1]]},"references-count":48,"journal-issue":{"issue":"1"},"URL":"https:\/\/doi.org\/10.4018\/ijmcmc.2020010102","relation":{},"ISSN":["1937-9412","1937-9404"],"issn-type":[{"value":"1937-9412","type":"print"},{"value":"1937-9404","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,1]]}}}