{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T00:31:17Z","timestamp":1770424277022,"version":"3.49.0"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T00:00:00Z","timestamp":1770336000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T00:00:00Z","timestamp":1770336000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-026-04770-7","type":"journal-article","created":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T12:30:53Z","timestamp":1770381053000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Using Reinforcement Learning to Boost Grey Wolf Optimizer for Influence Maximization in Social Networks"],"prefix":"10.1007","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9843-5886","authenticated-orcid":false,"given":"Mehdy","family":"Roayaei","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,6]]},"reference":[{"issue":"6","key":"4770_CR1","doi-asserted-by":"publisher","first-page":"3406","DOI":"10.1109\/TCSS.2022.3198096","volume":"10","author":"B Razaghi","year":"2022","unstructured":"Razaghi B, Roayaei M, Charkari NM. On the group-fairness-aware influence maximization in social networks. IEEE Trans Comput Soc Syst. 2022;10(6):3406\u201314.","journal-title":"IEEE Trans Comput Soc Syst"},{"key":"4770_CR2","doi-asserted-by":"crossref","unstructured":"Ali AF, Hassanien AE. A survey of metaheuristics methods for bioinformatics applications. In Applications of intelligent optimization in biology and medicine: Current trends and open problems. Cham: Springer International Publishing. 2015:23-46.","DOI":"10.1007\/978-3-319-21212-8_2"},{"key":"4770_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2023.108718","volume":"108","author":"N Alasmari","year":"2023","unstructured":"Alasmari N, Alohali MA, Khalid M, Almalki N, Motwakel A, Alsaid MI, et al. Improved metaheuristics with deep learning based object detector for intelligent control in autonomous vehicles. Comput Electr Eng Comput Electr Eng. 2023;108:108718.","journal-title":"Comput Electr Eng Comput Electr Eng"},{"issue":"6","key":"4770_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3459664","volume":"54","author":"EG Talbi","year":"2021","unstructured":"Talbi EG. Machine learning into metaheuristics: a survey and taxonomy. ACM Comput Surv (CSUR). 2021;54(6):1\u201332.","journal-title":"ACM Comput Surv (CSUR)"},{"key":"4770_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.cor.2021.105400","volume":"134","author":"N Mazyavkina","year":"2021","unstructured":"Mazyavkina N, Sviridov S, Ivanov S, Burnaev E. Reinforcement learning for combinatorial optimization: a survey. Comput Oper Res. 2021;134:105400.","journal-title":"Comput Oper Res"},{"key":"4770_CR6","doi-asserted-by":"crossref","unstructured":"Kempe D, Kleinberg J, Tardos \u00c9. Maximizing the spread of influence through a social network. In Proceedings of the ninth ACM SIGKDD international conference on knowledge discovery and data mining. 2003:137-146.","DOI":"10.1145\/956750.956769"},{"key":"4770_CR7","doi-asserted-by":"publisher","first-page":"360","DOI":"10.1016\/j.asoc.2018.02.016","volume":"66","author":"Q He","year":"2018","unstructured":"He Q, Wang X, Huang M, Lv J, Ma L. Heuristics-based influence maximization for opinion formation in social networks. Appl Soft Comput. 2018;66:360\u20139.","journal-title":"Appl Soft Comput"},{"key":"4770_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2019.112971","volume":"142","author":"A Zareie","year":"2020","unstructured":"Zareie A, Sheikhahmadi A, Jalili M. Identification of influential users in social network using gray wolf optimization algorithm. Expert Syst Appl. 2020;142:112971.","journal-title":"Expert Syst Appl"},{"key":"4770_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2020.106436","volume":"94","author":"A Sheikhahmadi","year":"2020","unstructured":"Sheikhahmadi A, Zareie A. Identifying influential spreaders using multi-objective artificial bee colony optimization. Appl Soft Comput. 2020;94:106436.","journal-title":"Appl Soft Comput"},{"key":"4770_CR10","doi-asserted-by":"publisher","first-page":"78","DOI":"10.1016\/j.ins.2021.01.013","volume":"562","author":"C Wang","year":"2021","unstructured":"Wang C, Wang H, Chen H, Li D. Attributed community search based on effective scoring function and elastic greedy method. Inf Sci. 2021;562:78\u201393.","journal-title":"Inf Sci"},{"issue":"6","key":"4770_CR11","doi-asserted-by":"publisher","first-page":"881","DOI":"10.1177\/1090198106297855","volume":"34","author":"TW Valente","year":"2007","unstructured":"Valente TW, Pumpuang P. Identifying opinion leaders to promote behavior change. Health Educ Behav. 2007;34(6):881\u201396.","journal-title":"Health Educ Behav"},{"key":"4770_CR12","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1016\/j.ins.2017.10.031","volume":"426","author":"Y Zhao","year":"2018","unstructured":"Zhao Y, Kou G, Peng Y, Chen Y. Understanding influence power of opinion leaders in e-commerce networks: an opinion dynamics theory perspective. Inf Sci. 2018;426:131\u201347.","journal-title":"Inf Sci"},{"key":"4770_CR13","doi-asserted-by":"crossref","unstructured":"Richardson M, Domingos P. Mining knowledge-sharing sites for viral marketing. In Proceedings of the eighth ACM SIGKDD international conference on knowledge discovery and data mining. 2002:61-70.","DOI":"10.1145\/775047.775057"},{"issue":"1","key":"4770_CR14","first-page":"591","volume":"26","author":"W Chen","year":"2012","unstructured":"Chen W, Lu W, Zhang N. Time-critical influence maximization in social networks with time-delayed diffusion process. Proceed AAAI Conf Artif Intell. 2012;26(1):591\u20138.","journal-title":"Proceed AAAI Conf Artif Intell"},{"key":"4770_CR15","first-page":"48","volume":"10","author":"M Rahimi","year":"2024","unstructured":"Rahimi M, Roayaei M. A multi-view rumor detection framework using dynamic propagation structure, interaction network, and content. IEEE Trans Signal Inf Process Netw. 2024;10:48\u201358.","journal-title":"IEEE Trans Signal Inf Process Netw"},{"key":"4770_CR16","first-page":"65211","volume":"11","author":"MN Makhadmeh","year":"2023","unstructured":"Makhadmeh MN, Al-Betar MA, Doush IA, Awadallah MA, Kassaymeh S, Mirjalili S, et al. Recent advances in Grey Wolf Optimizer, its versions and applications. IEEE Access. 2023;11:65211\u201330.","journal-title":"IEEE Access"},{"key":"4770_CR17","doi-asserted-by":"publisher","first-page":"46","DOI":"10.1016\/j.advengsoft.2013.12.007","volume":"69","author":"S Mirjalili","year":"2014","unstructured":"Mirjalili S, Mirjalili SM, Lewis A. Grey Wolf Optimizer. Adv Eng Softw. 2014;69:46\u201361.","journal-title":"Adv Eng Softw"},{"key":"4770_CR18","doi-asserted-by":"publisher","first-page":"371","DOI":"10.1016\/j.neucom.2015.06.083","volume":"172","author":"E Emary","year":"2016","unstructured":"Emary E, Zawbaa HM, Hassanien AE. Binary grey wolf optimization approaches for feature selection. Neurocomputing. 2016;172:371\u201381.","journal-title":"Neurocomputing"},{"key":"4770_CR19","doi-asserted-by":"crossref","unstructured":"Esmaili A, Roayaei M. UAV-Based warehouse management using multi-agent RL: Applications, challenges, and solutions. In Applications of Machine Learning in UAV Networks. IGI Global Scientific Publishing. 2024:263\u2013306.","DOI":"10.4018\/979-8-3693-0578-2.ch011"},{"key":"4770_CR20","doi-asserted-by":"crossref","unstructured":"Rezaei Gazik MA, Roayaei M. Batch (offline) reinforcement learning for recommender system. In 2023 31st International Conference on Electrical Engineering (ICEE). 2023:245\u2013250.","DOI":"10.1109\/ICEE59167.2023.10334722"},{"issue":"2","key":"4770_CR21","doi-asserted-by":"publisher","first-page":"1543","DOI":"10.1007\/s10462-022-10205-5","volume":"56","author":"V Uc-Cetina","year":"2023","unstructured":"Uc-Cetina V, Navarro-Guerrero N, Martin-Gonzalez A, Weber C, Wermter S. Survey on reinforcement learning for language processing. Artif Intell Rev. 2023;56(2):1543\u201375.","journal-title":"Artif Intell Rev"},{"key":"4770_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.artmed.2020.101964","volume":"109","author":"A Coronato","year":"2020","unstructured":"Coronato A, Naeem M, De Pietro G, Paragliola G. Reinforcement learning for intelligent healthcare applications: a survey. Artif Intell Med. 2020;109:101964.","journal-title":"Artif Intell Med"},{"key":"4770_CR23","first-page":"279","volume":"8","author":"C Watkins","year":"1992","unstructured":"Watkins C, Dayan P. Q-learning. Mach Learn. 1992;8:279\u201392.","journal-title":"Mach Learn"},{"issue":"7540","key":"4770_CR24","doi-asserted-by":"publisher","first-page":"529","DOI":"10.1038\/nature14236","volume":"518","author":"V Mnih","year":"2015","unstructured":"Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, et al. Human-level control through deep reinforcement learning. Nature. 2015;518(7540):529\u201333.","journal-title":"Nature"},{"key":"4770_CR25","unstructured":"Khalil E, Dai H, Zhang Y, Dilkina B, Song L. Learning combinatorial optimization algorithms over graphs. Adv Neural Inform Process Syst. 2017:30."},{"key":"4770_CR26","unstructured":"Abe K, Xu Z, Sato I, Sugiyama M. Solving NP-hard problems on graphs with extended AlphaGo Zero. arXiv preprint arXiv:1905.11623. 2019."},{"key":"4770_CR27","doi-asserted-by":"crossref","unstructured":"Deudon M, Cournut P, Lacoste, A, Adulyasak Y, Rousseau LM. Learning heuristics for the TSP by policy gradient. In Integration of Constraint Programming, Artificial Intelligence, and Operations Research: 15th International Conference, CPAIOR 2018, Delft, The Netherlands, June 26\u201329, 2018. 2018:170-181.","DOI":"10.1007\/978-3-319-93031-2_12"},{"issue":"2","key":"4770_CR28","doi-asserted-by":"publisher","first-page":"298","DOI":"10.3390\/math8020298","volume":"8","author":"S Gu","year":"2020","unstructured":"Gu S, Yang Y. A deep learning algorithm for the max-cut problem based on pointer network structure with supervised learning and reinforcement learning strategies. Mathematics. 2020;8(2):298.","journal-title":"Mathematics"},{"key":"4770_CR29","unstructured":"Nazari M, Oroojlooy A, Snyder L, Tak\u00e1c M. Reinforcement learning for solving the vehicle routing problem. Adv Neural Inform Process Syst. 2018:31."},{"key":"4770_CR30","unstructured":"Catteeuw D, Drugan M, Manderick B. \u2019Guided\u2019 Restarts Hill-Climbing. In In Search of Synergies Between Reinforcement Learning and Evolutionary Computation,... Workshop at the 13th International Conference on Parallel Problem Solving from Nature, Ljubljana, Slovenia. 2014:1-4."},{"issue":"1","key":"4770_CR31","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1109\/4235.585892","volume":"1","author":"M Dorigo","year":"1997","unstructured":"Dorigo M, Gambardella LM. Ant colony system: a cooperative learning approach to the traveling salesman problem. IEEE Trans Evol Comput. 1997;1(1):53\u201366.","journal-title":"IEEE Trans Evol Comput"},{"key":"4770_CR32","doi-asserted-by":"crossref","unstructured":"Seyyedabbasi A, Aliyev R, Kiani F, Gulle MU, Basyildiz H, Shah MA. Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems. Knowl Based Syst. 2021:223.","DOI":"10.1016\/j.knosys.2021.107044"},{"key":"4770_CR33","doi-asserted-by":"crossref","unstructured":"Chen Q, Huang M, Xu Q, Wang H, Wang J. Reinforcement learning-based genetic algorithm in optimizing multidimensional data discretization scheme. Math Probl Eng. 2020:1-13.","DOI":"10.1155\/2020\/1698323"},{"issue":"32","key":"4770_CR34","doi-asserted-by":"publisher","first-page":"10007","DOI":"10.1007\/s00521-019-04527-9","volume":"2020","author":"Y Xu","year":"2020","unstructured":"Xu Y, Pi D. A reinforcement learning-based communication topology in particle swarm optimization. Neural Comput Appl. 2020;2020(32):10007\u201332.","journal-title":"Neural Comput Appl"},{"issue":"2","key":"4770_CR35","doi-asserted-by":"publisher","first-page":"329","DOI":"10.1080\/0305215X.2020.1867120","volume":"54","author":"D Wu","year":"2022","unstructured":"Wu D, Wang GG. Employing reinforcement learning to enhance particle swarm optimization methods. Eng Optim. 2022;54(2):329\u201348.","journal-title":"Eng Optim"},{"key":"4770_CR36","first-page":"2021","volume-title":"21st International Conference","author":"M Becerra-Rozas","year":"2021","unstructured":"Becerra-Rozas M, Lemus-Romani J, Crawford B, Soto R, Cisternas-Caneo F, Embry AT, et al. 21st International Conference. Italy: Cagliari; 2021. p. 2021."},{"key":"4770_CR37","doi-asserted-by":"crossref","unstructured":"Zhao F, Hu X, Wang L, Zhao J, Tang J. A reinforcement learning brain storm optimization algorithm (BSO) with learning mechanism. Knowl Based Syst. 2022:235.","DOI":"10.1016\/j.knosys.2021.107645"},{"key":"4770_CR38","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1016\/j.swevo.2018.03.014","volume":"43","author":"AK Sadhu","year":"2018","unstructured":"Sadhu AK, Konar A, Bhattacharjee T, Das S. Synergism of firefly algorithm and Q-learning for robot arm path planning. Swarm Evol Comput. 2018;43:50\u201368.","journal-title":"Swarm Evol Comput"},{"key":"4770_CR39","doi-asserted-by":"crossref","unstructured":"Qu C, Gai W, Zhong M, Zhang J. A novel reinforcement learning based grey wolf optimizer algorithm for unmanned aerial vehicles (UAVs) path planning. Appl Soft Comput. 2020:89.","DOI":"10.1016\/j.asoc.2020.106099"},{"key":"4770_CR40","doi-asserted-by":"crossref","unstructured":"Chen R, Yang B, Li S, Wang S, Cheng Q. An effective multi-population grey wolf optimizer based on reinforcement learning for flow shop scheduling problem with multi-machine collaboration. Comput Ind Eng. 2021:162.","DOI":"10.1016\/j.cie.2021.107738"},{"key":"4770_CR41","doi-asserted-by":"crossref","unstructured":"Emary E, Zawbaa HM, Grosan C. Experienced gray wolf optimization through reinforcement learning and neural networks. IEEE Trans Neural Netw Learn Syst. 2018;29(3).","DOI":"10.1109\/TNNLS.2016.2634548"},{"key":"4770_CR42","doi-asserted-by":"crossref","unstructured":"Tapia D, Crawford B, Soto R, Palma W, Lemus-Romani J, Cisternas-Caneo F, Misra S. Embedding Q-learning in the selection of metaheuristic operators: The enhanced binary grey wolf optimizer case. In 2021 IEEE International Conference on Automation\/XXIV Congress of the Chilean Association of Automatic Control (ICA-ACCA). 2021:1-6.","DOI":"10.1109\/ICAACCA51523.2021.9465259"},{"key":"4770_CR43","doi-asserted-by":"publisher","first-page":"2403","DOI":"10.1007\/s11277-020-07981-0","volume":"117","author":"N Yuvaraj","year":"2021","unstructured":"Yuvaraj N, Karthikeyan T, Praghash K. An improved task allocation scheme in serverless computing using gray wolf optimization (GWO) based reinforcement learning (RIL) approach. Wirel Pers Commun. 2021;117:2403\u201321.","journal-title":"Wirel Pers Commun"},{"issue":"4","key":"4770_CR44","doi-asserted-by":"publisher","first-page":"3659","DOI":"10.1109\/TASE.2021.3129439","volume":"19","author":"C Lin","year":"2022","unstructured":"Lin C, Cao Z, Zhou M. Learning-based grey wolf optimizer for stochastic flexible job shop scheduling. IEEE Trans Autom Sci Eng. 2022;19(4):3659\u201371.","journal-title":"IEEE Trans Autom Sci Eng"},{"key":"4770_CR45","doi-asserted-by":"publisher","first-page":"10451","DOI":"10.1007\/s00521-019-04580-4","volume":"32","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Jin Z, Chen Y. Hybridizing grey wolf optimization with neural network algorithm for global numerical optimization problems. Neural Comput Appl. 2020;32:10451\u201370.","journal-title":"Neural Comput Appl"},{"key":"4770_CR46","doi-asserted-by":"crossref","unstructured":"Kumar R, Singh L, Tiwari R. Novel reinforcement learning guided enhanced variable weight grey wolf optimization (RLV-GWO) algorithm for multi-UAV path planning. Wirel Pers Commun. 2023:1-31.","DOI":"10.1007\/s11277-023-10534-w"},{"issue":"23","key":"4770_CR47","doi-asserted-by":"publisher","first-page":"14715","DOI":"10.1007\/s00500-021-06282-3","volume":"25","author":"M Roayaei","year":"2021","unstructured":"Roayaei M. On the binarization of grey wolf optimizer: a novel binary optimizer algorithm. Soft Comput. 2021;25(23):14715\u201328.","journal-title":"Soft Comput"},{"issue":"3","key":"4770_CR48","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1002\/sim.8803","volume":"40","author":"S Kim","year":"2021","unstructured":"Kim S, Lee S, Choi JI, Cho H. Optimal joinpoint detection using binary genetic algorithm: cancer trend analysis application. Stat Med. 2021;40(3):799\u2013822.","journal-title":"Stat Med"},{"issue":"2","key":"4770_CR49","doi-asserted-by":"publisher","first-page":"266","DOI":"10.3390\/biomimetics8020266","volume":"8","author":"J Zhu","year":"2023","unstructured":"Zhu J, Liu J, Chen Y, Xue X, Sun S. Binary restructuring particle swarm optimization and its application. Biomimetics. 2023;8(2):266.","journal-title":"Biomimetics"},{"key":"4770_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.114817","volume":"175","author":"MS Kiran","year":"2021","unstructured":"Kiran MS. Performance assessment of a binary artificial bee colony algorithm. Expert Syst Appl. 2021;175:114817.","journal-title":"Expert Syst Appl"},{"key":"4770_CR51","doi-asserted-by":"crossref","unstructured":"Saheed YK. Binary firefly algorithm-based feature selection method for high-dimensional intrusion detection data. In Illumination of Artificial Intelligence in Cybersecurity and Forensics. Cham: Springer International Publishing. 2022:273-288","DOI":"10.1007\/978-3-030-93453-8_12"},{"issue":"20","key":"4770_CR52","doi-asserted-by":"publisher","first-page":"2611","DOI":"10.3390\/math9202611","volume":"9","author":"J Garc\u00eda","year":"2021","unstructured":"Garc\u00eda J, Lemus-Romani J, Altimiras F, Crawford B, Soto R, Becerra-Rozas M, et al. A binary machine learning cuckoo search algorithm with local search operator for the set-union knapsack problem. Mathematics. 2021;9(20):2611.","journal-title":"Mathematics"},{"key":"4770_CR53","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107470","volume":"107","author":"RCT de Souza","year":"2020","unstructured":"de Souza RCT, de Macedo CA, dos Santos CL, Pierezan J, Mariani VC. Binary coyote optimization algorithm for feature selection. Pattern Recogn. 2020;107:107470.","journal-title":"Pattern Recogn"},{"key":"4770_CR54","doi-asserted-by":"crossref","unstructured":"Zareie A, Sheikhahmadi A, Jalili M. Identification of influential users in social network using gray wolf optimization algorithm. Expert Syst Appl. 2020:142.","DOI":"10.1016\/j.eswa.2019.112971"},{"key":"4770_CR55","doi-asserted-by":"crossref","unstructured":"Beni HA, Bouyer A. TI-SC: Top-k influential nodes selection based on community detection and scoring criteria in social networks. J Ambient Intell Humanized Comput. 2020:1-20.","DOI":"10.1007\/s12652-020-01760-2"},{"key":"4770_CR56","unstructured":"Roayaei M. RLSetGWO. GitHub Repository. Available at: https:\/\/github.com\/mroayaei\/RLSetGWO.git. Accessed 5 Sep 2024."},{"key":"4770_CR57","doi-asserted-by":"crossref","unstructured":"Geng Y, Wang K, Liu Z, Yu M, Yu JX. Influence maximization revisited. In Australasian Database Conference, 2023:356-370.","DOI":"10.1007\/978-3-031-47843-7_25"},{"key":"4770_CR58","doi-asserted-by":"crossref","unstructured":"Ghalghachi AM, Ardakani MR. Finding influential individuals in social networks using deep reinforcement learning to tune the neighborhood selection operator in the variable neighborhood search algorithm. In 2024 10th International Conference on Web Research (ICWR). 2024:112-118.","DOI":"10.1109\/ICWR61162.2024.10533378"},{"key":"4770_CR59","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1613\/jair.1.12997","volume":"74","author":"W Chen","year":"2022","unstructured":"Chen W, Peng B, Schoenebeck G, Tao B. Adaptive greedy versus non-adaptive greedy for influence maximization. J Artif Intell Res. 2022;74:303\u201351.","journal-title":"J Artif Intell Res"},{"key":"4770_CR60","doi-asserted-by":"publisher","DOI":"10.1016\/j.swevo.2023.101387","volume":"83","author":"Z Hu","year":"2023","unstructured":"Hu Z, Gong W, Pedrycz W, Li Y. Deep reinforcement learning assisted co-evolutionary differential evolution for constrained optimization. Swarm Evol Comput. 2023;83:101387.","journal-title":"Swarm Evol Comput"},{"key":"4770_CR61","doi-asserted-by":"publisher","first-page":"688","DOI":"10.1016\/j.applthermaleng.2018.10.020","volume":"146","author":"TC Bora","year":"2019","unstructured":"Bora TC, Mariani VC, dos Santos CL. Multi-objective optimization of the environmental-economic dispatch with reinforcement learning based on non-dominated sorting genetic algorithm. Appl Therm Eng. 2019;146:688\u2013700.","journal-title":"Appl Therm Eng"},{"issue":"5","key":"4770_CR62","doi-asserted-by":"publisher","first-page":"1260","DOI":"10.1109\/TEVC.2022.3199045","volume":"27","author":"W Liu","year":"2022","unstructured":"Liu W, Wang R, Zhang T, Li K, Li W, Ishibuchi H, et al. Hybridization of evolutionary algorithm and deep reinforcement learning for multiobjective orienteering optimization. IEEE Trans Evol Comput. 2022;27(5):1260\u201374.","journal-title":"IEEE Trans Evol Comput"}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-026-04770-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-026-04770-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-026-04770-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T12:30:59Z","timestamp":1770381059000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-026-04770-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,6]]},"references-count":62,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["4770"],"URL":"https:\/\/doi.org\/10.1007\/s42979-026-04770-7","relation":{},"ISSN":["2661-8907"],"issn-type":[{"value":"2661-8907","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,6]]},"assertion":[{"value":"6 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"184"}}