{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T18:26:29Z","timestamp":1770747989765,"version":"3.49.0"},"reference-count":87,"publisher":"Maximum Academic Press","license":[{"start":{"date-parts":[[2018,11,21]],"date-time":"2018-11-21T00:00:00Z","timestamp":1542758400000},"content-version":"unspecified","delay-in-days":324,"URL":"https:\/\/www.cambridge.org\/core\/terms"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["The Knowledge Engineering Review"],"published-print":{"date-parts":[[2018]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>Automated planning has been a continuous field of study since the 1960s, since the notion of accomplishing a task using an ordered set of actions resonates with almost every known activity domain. However, as we move from toy domains closer to the complex real world, these actions become increasingly difficult to codify. The reasons range from intense laborious effort, to intricacies so barely identifiable, that programming them is a challenge that presents itself much later in the process. In such domains, planners now leverage recent advancements in machine learning to learn action models, that is, blueprints of all the actions whose execution effectuates transitions in the system. This learning provides an opportunity for the evolution of the model toward a version more consistent and adapted to its environment, augmenting the probability of success of the plans. It is also a conscious effort to decrease laborious manual coding and increase quality. This paper presents a survey of the machine learning techniques applied for learning planning action models. It first describes the characteristics of learning systems. It then details the learning techniques that have been used in the literature during the past decades, and finally presents some open issues.<\/jats:p>","DOI":"10.1017\/s0269888918000188","type":"journal-article","created":{"date-parts":[[2018,11,21]],"date-time":"2018-11-21T04:23:28Z","timestamp":1542774208000},"source":"Crossref","is-referenced-by-count":34,"title":["A review of learning planning action models"],"prefix":"10.48130","volume":"33","author":[{"given":"Ankuj","family":"Arora","sequence":"first","affiliation":[]},{"given":"Humbert","family":"Fiorino","sequence":"additional","affiliation":[]},{"given":"Damien","family":"Pellier","sequence":"additional","affiliation":[]},{"given":"Marc","family":"M\u00e9tivier","sequence":"additional","affiliation":[]},{"given":"Sylvie","family":"Pesty","sequence":"additional","affiliation":[]}],"member":"27968","published-online":{"date-parts":[[2018,11,21]]},"reference":[{"key":"S0269888918000188_ref99","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2014.03.004"},{"key":"S0269888918000188_ref55","unstructured":"Mour\u00e3o K. , Petrick R. P. & Steedman M. 2008. Using kernel perceptrons to learn action effects for planning. In International Conference on Cognitive Systems (CogSys 2008), 45\u201350."},{"key":"S0269888918000188_ref18","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-008-9045-x"},{"key":"S0269888918000188_ref79","first-page":"53","article-title":"Knowledge engineering tools in planning: state-of-the-art and future challenges","volume":"53","author":"Shah","year":"2013","journal-title":"Knowledge Engineering for Planning and Scheduling"},{"key":"S0269888918000188_ref91","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2006.11.005"},{"key":"S0269888918000188_ref27","unstructured":"Ferrer-Mestres J. , Frances G. & Geffner H. 2015. Planning with state constraints and its application to combined task and motion planning. In Proceedings of Workshop on Planning and Robotics (PLANROB), 13\u201322."},{"key":"S0269888918000188_ref42","unstructured":"Jim\u00e9nez S. , Fern\u00e1ndez F. & Borrajo D. 2008. The PELA architecture: integrating planning and learning to improve execution. In Association for the Advancement of Artificial Intelligence."},{"key":"S0269888918000188_ref10","unstructured":"Cashmore M. , Fox M. , Long D. , Ridder B. C. & Magazzeni D. 2016b. Strategic planning for autonomous systems over long horizons. In Proceedings of the 4th ICAPS Workshop on Planning and Robotics (PlanRob 2016), 74\u201381."},{"key":"S0269888918000188_ref46","doi-asserted-by":"crossref","unstructured":"Mart\u00ednez D. , Alenya G. , Torras C. , Ribeiro T. & Inoue K. 2016. Learning relational dynamics of stochastic domains for planning. In Proceedings of the 26th International Conference on Automated Planning and Scheduling.","DOI":"10.1609\/icaps.v26i1.13746"},{"key":"S0269888918000188_ref66","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2009.191"},{"key":"S0269888918000188_ref33","unstructured":"Gregory J. , Fink J. , Rogers J. , Gupta S. & Crowley J. L. 2016. A risk-based framework for incorporating navigation uncertainty into exploration strategies. In Proceedings of the 4th ICAPS Workshop on Planning and Robotics (PlanRob 2016), 176\u2013183."},{"key":"S0269888918000188_ref14","doi-asserted-by":"crossref","unstructured":"Cresswell S. & Gregory P. 2011. Generalised Domain Model Acquisition from Action Traces. In International Conference on Automated Planning and Scheduling.","DOI":"10.1609\/icaps.v21i1.13476"},{"key":"S0269888918000188_ref15","doi-asserted-by":"crossref","unstructured":"Cresswell S. N. , McCluskey T. L. & West M. M. 2009. Acquisition of object-centered domain models from planning examples.","DOI":"10.1609\/icaps.v19i1.13391"},{"key":"S0269888918000188_ref93","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1880"},{"key":"S0269888918000188_ref67","unstructured":"Pasula H. , Zettlemoyer L. S. & Kaelbling L. P. 2004. Learning probabilistic relational planning rules. In International Conference on Automated Planning and Scheduling, 73\u201382."},{"key":"S0269888918000188_ref30","unstructured":"Gil Y. 1992. Acquiring Domain Knowledge for Planning by Experimentation. PhD thesis, Department of Computer Science, Carnegie-Mellon University."},{"key":"S0269888918000188_ref81","doi-asserted-by":"publisher","DOI":"10.1038\/nature16961"},{"key":"S0269888918000188_ref8","first-page":"75","article-title":"Mixed-initiative planning in space mission operations","volume":"28","author":"Bresina","year":"2007","journal-title":"AI Magazine"},{"key":"S0269888918000188_ref80","doi-asserted-by":"publisher","DOI":"10.1007\/BF00993064"},{"key":"S0269888918000188_ref69","first-page":"324","article-title":"ADL: exploring the middle ground between STRIPS and the situation calculus","volume":"89","author":"Pednault","year":"1989","journal-title":"Kr"},{"key":"S0269888918000188_ref74","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-006-5833-1"},{"key":"S0269888918000188_ref37","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-013-5353-8"},{"key":"S0269888918000188_ref5","unstructured":"Bevacqua G. , Cacace J. , Finzi A. & Lippiello V. 2015. Mixed-initiative planning and execution for multiple drones in search and rescue missions. In ICAPS, pp. 315\u2013323."},{"key":"S0269888918000188_ref97","unstructured":"Zhuo H. H. , Mu\u00f1oz-Avila H. & Yang Q. 2011. Learning action models for multi-agent planning. In The 10th International Conference on Autonomous Agents and Multiagent Systems, 1, 217\u2013224."},{"key":"S0269888918000188_ref71","doi-asserted-by":"publisher","DOI":"10.1007\/BF00116251"},{"key":"S0269888918000188_ref98","unstructured":"Zhuo H. H. , Nguyen T. A. & Kambhampati S. 2013. Refining incomplete planning domain models through plan traces. In International Joint Conference on Artificial Intelligence."},{"key":"S0269888918000188_ref16","doi-asserted-by":"publisher","DOI":"10.1017\/S0269888912000422"},{"key":"S0269888918000188_ref75","unstructured":"Rodrigues C. , G\u00e9rard P. , Rouveirol C. & Soldano H. 2011. Active learning of relational action models. In International Conference on Inductive Logic Programming, 302\u2013316."},{"key":"S0269888918000188_ref59","doi-asserted-by":"publisher","DOI":"10.1016\/S0004-3702(98)00068-X"},{"key":"S0269888918000188_ref63","volume-title":"Shakey the Robot","author":"Nilsson","year":"1984"},{"key":"S0269888918000188_ref53","unstructured":"Mour\u00e3o K. 2012. Learning action representations using kernel perceptrons."},{"key":"S0269888918000188_ref88","unstructured":"Weber B. G. , Mateas M. & Jhala A. 2012. Learning from demonstration for goal-driven autonomy. In Association for the Advancement of Artificial Intelligence."},{"key":"S0269888918000188_ref21","unstructured":"Driessens K. , Ramon J. & Blockeel H. 2001. Speeding up relational reinforcement learning through the use of an incremental first order decision tree learner. In European Conference on Machine Learning, 97\u2013108."},{"key":"S0269888918000188_ref86","unstructured":"Walsh T. J. & Littman M. L. 2008. Efficient learning of action schemas and web-service descriptions. In Association for the Advancement of Artificial Intelligence, 714\u2013719."},{"key":"S0269888918000188_ref100","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2010.09.007"},{"key":"S0269888918000188_ref4","unstructured":"Balac N. , Gaines D. M. & Fisher D. 2000. Learning action models for navigation in noisy environments. In ICML Workshop on Machine Learning of Spatial Knowledge."},{"key":"S0269888918000188_ref6","unstructured":"Brafman R. I. & Domshlak C. 2008. From one to many: planning for loosely coupled multi-agent systems. In ICAPS,28\u201335."},{"key":"S0269888918000188_ref9","unstructured":"Cashmore M. , Fox M. , Long D. , Ridder B. C. & Magazzeni D. 2016a. Opportunistic planning for increased plan utility. In Proceedings of the 4th ICAPS Workshop on Planning and Robotics (PlanRob 2016), 82\u201392."},{"key":"S0269888918000188_ref85","volume-title":"Reinforcement Learning: An Introduction","author":"Sutton","year":"1998"},{"key":"S0269888918000188_ref12","unstructured":"Cohen L. , Shimony S. E. & Weiss G. 2015. Estimating the probability of meeting a deadline in hierarchical plans. In Computational Logic in Multi-Agent Systems, 243\u2013258."},{"key":"S0269888918000188_ref17","unstructured":"Croonenborghs T. , Ramon J. , Blockeel H. & Bruynooghe M. 2007. Online learning and exploiting relational models in reinforcement learning. In International Joint Conference on Artificial Intelligence, 726\u2013731."},{"key":"S0269888918000188_ref87","unstructured":"Wang X. 1996. Learning Planning Operators by Observation and Practice. Doctoral dissertation, Carnegie Mellon University."},{"key":"S0269888918000188_ref19","unstructured":"Deshpande A. , Milch B. , Zettlemoyer L. S. & Kaelbling L. 2007. Learning probabilistic relational dynamics for multiple tasks. In Probabilistic, Logical and Relational Learning - A Further Synthesis."},{"key":"S0269888918000188_ref20","first-page":"36","article-title":"Closed loop configuration planning with time and resources","author":"Di Rocco","year":"2013","journal-title":"Planning and Robotics"},{"key":"S0269888918000188_ref22","doi-asserted-by":"publisher","DOI":"10.1023\/A:1007694015589"},{"key":"S0269888918000188_ref24","first-page":"3","article-title":"Theoretical underpinnings of situation awareness: a critical review","volume":"1","author":"Endsley","year":"2000","journal-title":"Situation Awareness Analysis and Measurement"},{"key":"S0269888918000188_ref26","doi-asserted-by":"publisher","DOI":"10.1016\/0004-3702(71)90010-5"},{"key":"S0269888918000188_ref28","doi-asserted-by":"publisher","DOI":"10.1023\/A:1008134010576"},{"key":"S0269888918000188_ref29","doi-asserted-by":"publisher","DOI":"10.1016\/B978-155860856-6\/50020-X"},{"key":"S0269888918000188_ref31","unstructured":"Guillame-Bert M. & Crowley J. L. 2012. Learning temporal association rules on symbolic time sequences. In Asian Conference on Machine Learning, 159\u2013174."},{"key":"S0269888918000188_ref32","unstructured":"Gregory P. & Cresswell S. 2015. Domain model acquisition in the presence of static relations in the LOP system. In ICAPS, 97\u2013105."},{"key":"S0269888918000188_ref34","doi-asserted-by":"crossref","unstructured":"Gregory P. & Lindsay A. 2016. Domain model acquisition in domains with action costs. In Proceedings of the Twenty-Sixth International Conference on International Conference on Automated Planning and Scheduling,149\u2013157. AAAI Press.","DOI":"10.1609\/icaps.v26i1.13762"},{"key":"S0269888918000188_ref77","unstructured":"Safaei J. & Ghassem-Sani G. 2007. Incremental learning of planning operators in stochastic domains. In International Conference on Current Trends in Theory and Practice of Computer Science, 644\u2013655."},{"key":"S0269888918000188_ref38","unstructured":"Jaidee U. , Mu\u00f1oz-Avila H. & Aha D. W. 2011. Integrated learning for goal-driven autonomy. In IJCAI, 22 (3): 2450."},{"key":"S0269888918000188_ref39","unstructured":"Jilani R. , Crampton A. , Kitchin D. E. & Vallati M. 2014. Automated knowledge engineering tools in planning: state-of-the-art and future challenges. In Knowledge Engineering for Planning and Scheduling."},{"key":"S0269888918000188_ref40","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24309-2_33"},{"key":"S0269888918000188_ref41","doi-asserted-by":"publisher","DOI":"10.1017\/S026988891200001X"},{"key":"S0269888918000188_ref44","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-010-9144-3"},{"key":"S0269888918000188_ref47","unstructured":"McCluskey T. L. , Cresswell S. N. , Richardson N. E. & West M. M. 2009. Automated acquisition of action knowledge."},{"key":"S0269888918000188_ref48","unstructured":"McCluskey T. L. , Richardson N. E. & Simpson R. M. 2002. An interactive method for inducing operator descriptions. In Artificial Intelligence Planning Systems, 121\u2013130."},{"key":"S0269888918000188_ref49","unstructured":"McDermott D. , Ghallab M. , Howe A. , Knoblock C. , Ram A. , Veloso M. , Weld D. & Wilkins D. 1998. PDDL-the planning domain definition language."},{"key":"S0269888918000188_ref50","doi-asserted-by":"publisher","DOI":"10.1613\/jair.4339"},{"key":"S0269888918000188_ref51","unstructured":"Molineaux M. & Aha D. W. 2014. Learning unknown event models. In Association for the Advancement of Artificial Intelligence, 395\u2013401."},{"key":"S0269888918000188_ref52","doi-asserted-by":"crossref","unstructured":"Molineaux M. , Klenk M. & Aha D. W. 2010. Goal-Driven Autonomy in a Navy Strategy Simulation. Knexus Research Corp.","DOI":"10.1609\/aaai.v24i1.7576"},{"key":"S0269888918000188_ref54","unstructured":"Mour\u00e3o K. 2014. Learning probabilistic planning operators from noisy observations. In Proceedings of the Workshop of the UK Planning and Scheduling Special Interest Group."},{"key":"S0269888918000188_ref56","unstructured":"Mour\u00e3o K. , Petrick R. P. & Steedman M. 2010. Learning action effects in partially observable domains. In European Conference on Artificial Intelligence, 973\u2013974."},{"key":"S0269888918000188_ref57","unstructured":"Mour\u00e3o K. , Zettlemoyer L. S. , Petrick R. & Steedman M. 2012. Learning STRIPS Operators from Noisy and Incomplete Observations. In Proceedings of the Twenty-Eighth Conference on Uncertainty in Artificial Intelligence, 614\u2013623."},{"key":"S0269888918000188_ref60","unstructured":"Newton M. A. H. & Levine J. 2010. Implicit learning of compiled macro-actions for planning. In European Conference on Artificial Intelligence, 323\u2013328."},{"key":"S0269888918000188_ref61","unstructured":"Newton M. A. H. , Levine J. , Fox M. & Long D. 2007. Learning macro-actions for arbitrary planners and domains. In International Conference on Automated Planning and Scheduling, 256\u2013263."},{"key":"S0269888918000188_ref62","unstructured":"Newton M. H. , Levine J. , Fox M. & Long D. 2008. Learning macros that are not captured by given example plans. In Poster Papers at the International Conference on Automated Planning and Scheduling."},{"key":"S0269888918000188_ref65","doi-asserted-by":"publisher","DOI":"10.1155\/2013\/942347"},{"key":"S0269888918000188_ref68","unstructured":"Pasula H. , Zettlemoyer L. S. & Kaelbling L. P. 2007. Learning symbolic models of stochastic domains. In Journal of Artificial Intelligence Research, 309\u2013352."},{"key":"S0269888918000188_ref70","unstructured":"Pell B. , Gat E. , Keesing R. , Muscettola N. & Smith B. 1997. Robust periodic planning and execution for autonomous spacecraft. In IJCAI, 1234\u20131239."},{"key":"S0269888918000188_ref73","unstructured":"Ranasinghe N. & Shen W. 2008. Surprise-based learning for developmental robotics. In Learning and Adaptive Behaviors for Robotic Systems, 65\u201370."},{"key":"S0269888918000188_ref76","unstructured":"Sadohara K. 2001. Learning of boolean functions using support vector machines. In International Conference on Algorithmic Learning Theory, 106\u2013118."},{"key":"S0269888918000188_ref78","unstructured":"Sanner S. 2010. Relational Dynamic Influence Diagram Language (rddl): Language Description. Unpublished Manuscript, Australian National University."},{"key":"S0269888918000188_ref82","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/615"},{"key":"S0269888918000188_ref83","unstructured":"Strenzke R. & Schulte A. 2011. The MMP: A mixed-initiative mission planning system for the multi-aircraft domain. In Proceeding of the International Conference on Automated Planning and Scheduling, 74\u201382."},{"key":"S0269888918000188_ref84","doi-asserted-by":"publisher","DOI":"10.1613\/jair.3451"},{"key":"S0269888918000188_ref89","doi-asserted-by":"publisher","DOI":"10.1016\/0010-0285(72)90002-3"},{"key":"S0269888918000188_ref92","unstructured":"Yoon S. & Kambhampati S. 2007. Towards model-lite planning: a proposal for learning and planning with incomplete domain models. In ICAPS 2007 Workshop on Artificial Intelligence Planning and Learning."},{"key":"S0269888918000188_ref94","unstructured":"Zettlemoyer L. S. , Pasula H. & Kaelbling L. P. 2005. Learning planning rules in noisy stochastic worlds. In Association for the Advancement of Artificial Intelligence, 911\u2013918."},{"key":"S0269888918000188_ref95","unstructured":"Zhang Y. , Sreedharan S. & Kambhampati S. 2015. Capability models and their applications in planning. In Proceedings of the 2015 International Conference on Autonomous Agents and Multiagent Systems, 1151\u20131159."},{"key":"S0269888918000188_ref96","unstructured":"Zhuo H. H. & Kambhampati S. 2013. Action-model acquisition from noisy plan traces. In International Joint Conference on Artificial Intelligence."},{"key":"S0269888918000188_ref103","first-page":"73","article-title":"Learning-assisted automated planning: looking back, taking stock, going forward","volume":"24","author":"Zimmerman","year":"2003","journal-title":"AI Magazine"},{"key":"S0269888918000188_ref58","unstructured":"Mu\u00f1oz-Avila H. , Aha D. W. , Breslow L. & Nau D. 1999. HICAP: an interactive case-based planning architecture and its application to noncombatant evacuation operations. In Proceedings of the Sixteenth National Conference on Artificial Intelligence and the Eleventh Innovative Applications of Artificial Intelligence Conference Innovative Applications of Artificial Intelligence, 870\u2013875."},{"key":"S0269888918000188_ref1","unstructured":"Agrawal R. & Srikant R. 1994. Fast algorithms for mining association rules. In Proceedings of the 20th International Conference on Very Large Data Bases, 487\u2013499."}],"container-title":["The Knowledge Engineering Review"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.cambridge.org\/core\/services\/aop-cambridge-core\/content\/view\/S0269888918000188","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,5]],"date-time":"2026-01-05T14:42:10Z","timestamp":1767624130000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.cambridge.org\/core\/product\/identifier\/S0269888918000188\/type\/journal_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018]]},"references-count":87,"alternative-id":["S0269888918000188"],"URL":"https:\/\/doi.org\/10.1017\/s0269888918000188","relation":{},"ISSN":["0269-8889","1469-8005"],"issn-type":[{"value":"0269-8889","type":"print"},{"value":"1469-8005","type":"electronic"}],"subject":[],"published":{"date-parts":[[2018]]},"article-number":"e20"}}