{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T23:14:05Z","timestamp":1740179645655,"version":"3.37.3"},"reference-count":56,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2018,6,1]],"date-time":"2018-06-01T00:00:00Z","timestamp":1527811200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"}],"funder":[{"name":"National Science and Engineering Research Council (Canada)","award":["RGPIN-2015-06117"],"award-info":[{"award-number":["RGPIN-2015-06117"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Games"],"published-print":{"date-parts":[[2018,6]]},"DOI":"10.1109\/tciaig.2017.2766980","type":"journal-article","created":{"date-parts":[[2017,10,26]],"date-time":"2017-10-26T19:50:06Z","timestamp":1509047406000},"page":"195-208","source":"Crossref","is-referenced-by-count":11,"title":["Discovering Agent Behaviors Through Code Reuse: Examples From Half-Field Offense and Ms. Pac-Man"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6071-4705","authenticated-orcid":false,"given":"Stephen","family":"Kelly","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1521-0671","authenticated-orcid":false,"given":"Malcolm I.","family":"Heywood","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1177\/105971230501300301"},{"key":"ref38","first-page":"537","article-title":"Scaling reinforcement learning toward RoboCup soccer","author":"stone","year":"2001","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICMLA.2007.23"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2003.819263"},{"key":"ref31","first-page":"234","article-title":"When novelty is not enough","volume":"6624","author":"cuccu","year":"2011","journal-title":"Appl Evol Comput"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1162\/evco.1997.5.4.373"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1023\/A:1025124423708"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pcbi.1000206"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.0611630104"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1162\/EVCO_a_00048"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/2001576.2001765"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/1830483.1830640"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-12239-2_6"},{"key":"ref2","doi-asserted-by":"crossref","first-page":"253","DOI":"10.1613\/jair.3912","article-title":"The arcade learning environment: An\n evaluation platform for general agents","volume":"47","author":"bellemare","year":"2012","journal-title":"J Artif Intell Res"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TCIAIG.2014.2339221"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1162\/106365600568086"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.2307\/2410639"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pcbi.1000719"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2006.887351"},{"key":"ref23","first-page":"937","article-title":"Accelerated neural evolution through cooperatively coevolved synapses","volume":"9","author":"gomez","year":"2008","journal-title":"J Mach Learn Res"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/1276958.1277293"},{"journal-title":"Linear Genetic Programming","year":"2007","author":"brameier","key":"ref25"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1145\/2330163.2330178"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CEC.2012.6252966"},{"key":"ref56","first-page":"1","article-title":"Statistical comparisons of classifiers over multiple data sets","volume":"7","author":"dem\u0161ar","year":"2006","journal-title":"J Mach Learn Res"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/TCIAIG.2013.2294713"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/CIG.2013.6633639"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1145\/1569901.1569918"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1177\/105971239700500305"},{"key":"ref11","doi-asserted-by":"crossref","DOI":"10.7551\/mitpress\/4151.001.0001","author":"stone","year":"2000","journal-title":"Layered Learning in Multiagent Systems"},{"key":"ref40","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1007\/11780519_9","article-title":"Keepaway soccer: From machine learning\n testbed to benchmark","author":"stone","year":"2006","journal-title":"RoboCup 2005 Robot Soccer World Cup IX"},{"key":"ref12","first-page":"64","article-title":"Behavior chaining&#x2014;Incremental behavior integration for\n evolutionary robotics","author":"bongard","year":"2008","journal-title":"Proc Artif Life 11th Int Conf Synth Simul Living Syst"},{"key":"ref13","first-page":"1543","article-title":"Evolving\n controllers for simulated car racing using object orientated genetic programming","author":"agapitos","year":"2007","journal-title":"Proc Genetic Evol Comput Conf"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-01181-8_26"},{"key":"ref15","article-title":"Empirical comparison of incremental reuse\n strategies in genetic programming for Keepaway soccer","author":"hsu","year":"2004","journal-title":"Proc Genetic Evol Comput Conf Late Breaking Papers"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-005-0460-9"},{"key":"ref17","article-title":"Transfer via\n inter-task mappings in policy search reinforcement learning","author":"taylor","year":"2007","journal-title":"Proc ACM Int Joint Conf Auton Agents Ultiagent Syst"},{"key":"ref18","first-page":"283","article-title":"Autonomous\n transfer for reinforcement learning","author":"taylor","year":"2008","journal-title":"Proc Int Joint Conf Auton Agents Multiagent Syst"},{"key":"ref19","first-page":"1737","article-title":"Evolving static representations for task transfer","volume":"11","author":"verbancsics","year":"2010","journal-title":"J Mach Learn Res"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1609\/aimag.v32i1.2329"},{"journal-title":"Reinforcement Learning An Introduction","year":"1998","author":"sutton","key":"ref3"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1162\/EVCO_a_00025"},{"key":"ref5","first-page":"1633","article-title":"Transfer learning for reinforcement learning domains: A survey","volume":"10","author":"taylor","year":"2009","journal-title":"J Mach Learn Res"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TCIAIG.2015.2390615"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CEC.2013.6557731"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4419-7747-2_3"},{"key":"ref9","doi-asserted-by":"crossref","first-page":"1143","DOI":"10.1145\/2739480.2754798","article-title":"Knowledge transfer from Keepaway soccer to half-field offense through program symbiosis:\n Building simple programs for a complex task","author":"kelly","year":"2015","journal-title":"Proc Genetic Evol Comput Conf"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/CIG.2011.6031997"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CIG.2012.6374165"},{"key":"ref48","first-page":"3110","article-title":"Genotypic versus behavioural diversity for teams of programs under the 4-v-3 Keepaway soccer\n task","author":"kelly","year":"2014","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CIG.2012.6374163"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-009-9100-2"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-74024-7_7"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-662-44303-3_7"},{"key":"ref43","first-page":"749","article-title":"An empirical analysis of value function-based and policy search reinforcement learning","author":"kalyanakrishnan","year":"2009","journal-title":"Proc 1st Int Conf Autonomous Agents Multiagent Syst"}],"container-title":["IEEE Transactions on Games"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7782673\/8386463\/08085186.pdf?arnumber=8085186","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,26]],"date-time":"2022-01-26T06:28:27Z","timestamp":1643178507000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8085186\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,6]]},"references-count":56,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/tciaig.2017.2766980","relation":{},"ISSN":["2475-1502","2475-1510"],"issn-type":[{"type":"print","value":"2475-1502"},{"type":"electronic","value":"2475-1510"}],"subject":[],"published":{"date-parts":[[2018,6]]}}}