{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,20]],"date-time":"2025-07-20T03:43:58Z","timestamp":1752983038937,"version":"3.40.3"},"publisher-location":"Cham","reference-count":35,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031514968"},{"type":"electronic","value":"9783031514975"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-51497-5_34","type":"book-chapter","created":{"date-parts":[[2024,1,31]],"date-time":"2024-01-31T09:16:09Z","timestamp":1706692569000},"page":"479-493","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["A Constrained-Optimization Approach to\u00a0the\u00a0Execution of\u00a0Prioritized Stacks of\u00a0Learned Multi-robot Tasks"],"prefix":"10.1007","author":[{"given":"Gennaro","family":"Notomista","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,1]]},"reference":[{"issue":"5","key":"34_CR1","doi-asserted-by":"publisher","first-page":"985","DOI":"10.1109\/TRO.2009.2017135","volume":"25","author":"G Antonelli","year":"2009","unstructured":"Antonelli, G.: Stability analysis for prioritized closed-loop inverse kinematic algorithms for redundant robotic systems. IEEE Trans. Rob. 25(5), 985\u2013994 (2009). https:\/\/doi.org\/10.1109\/TRO.2009.2017135","journal-title":"IEEE Trans. Rob."},{"key":"34_CR2","volume-title":"Reinforcement Learning and Optimal Control","author":"DP Bertsekas","year":"2019","unstructured":"Bertsekas, D.P.: Reinforcement Learning and Optimal Control. Athena Scientific Belmont, MA (2019)"},{"key":"34_CR3","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511804441","volume-title":"Convex Optimization","author":"S Boyd","year":"2004","unstructured":"Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge (2004)"},{"key":"34_CR4","doi-asserted-by":"publisher","DOI":"10.1201\/9781315137667","volume-title":"Applied Optimal Control: Optimization, Estimation, and Control","author":"AE Bryson","year":"2018","unstructured":"Bryson, A.E., Ho, Y.C.: Applied Optimal Control: Optimization, Estimation, and Control. Routledge, New York (2018)"},{"doi-asserted-by":"crossref","unstructured":"Bylard, A., Bonalli, R., Pavone, M.: Composable geometric motion policies using multi-task pullback bundle dynamical systems. arXiv preprint arXiv:2101.01297 (2021)","key":"34_CR5","DOI":"10.1109\/ICRA48506.2021.9561320"},{"unstructured":"Dulac-Arnold, G., Mankowitz, D., Hester, T.: Challenges of real-world reinforcement learning. arXiv preprint arXiv:1904.12901 (2019)","key":"34_CR6"},{"doi-asserted-by":"crossref","unstructured":"Freeman, R.A., Primbs, J.A.: Control Lyapunov functions: new ideas from an old source. In: Proceedings of 35th IEEE Conference on Decision and Control, vol.\u00a04, pp. 3926\u20133931. IEEE (1996)","key":"34_CR7","DOI":"10.1109\/CDC.1996.577294"},{"issue":"9","key":"34_CR8","doi-asserted-by":"publisher","first-page":"939","DOI":"10.1177\/0278364904045564","volume":"23","author":"BP Gerkey","year":"2004","unstructured":"Gerkey, B.P., Matari\u0107, M.J.: A formal analysis and taxonomy of task allocation in multi-robot systems. Int. J. Robot. Res. 23(9), 939\u2013954 (2004)","journal-title":"Int. J. Robot. Res."},{"unstructured":"Ghosh, D., Singh, A., Rajeswaran, A., Kumar, V., Levine, S.: Divide-and-conquer reinforcement learning. arXiv preprint arXiv:1711.09874 (2017)","key":"34_CR9"},{"doi-asserted-by":"crossref","unstructured":"Gupta, A., et al.: Reset-free reinforcement learning via multi-task learning: learning dexterous manipulation behaviors without human intervention. arXiv preprint arXiv:2104.11203 (2021)","key":"34_CR10","DOI":"10.1109\/ICRA48506.2021.9561384"},{"doi-asserted-by":"crossref","unstructured":"Haarnoja, T., Pong, V., Zhou, A., Dalal, M., Abbeel, P., Levine, S.: Composable deep reinforcement learning for robotic manipulation. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 6244\u20136251. IEEE (2018)","key":"34_CR11","DOI":"10.1109\/ICRA.2018.8460756"},{"unstructured":"Haarnoja, T., Tang, H., Abbeel, P., Levine, S.: Reinforcement learning with deep energy-based policies. In: International Conference on Machine Learning, pp. 1352\u20131361. PMLR (2017)","key":"34_CR12"},{"issue":"6506","key":"34_CR13","doi-asserted-by":"publisher","first-page":"915","DOI":"10.1126\/science.aaz7597","volume":"369","author":"LP Kaelbling","year":"2020","unstructured":"Kaelbling, L.P.: The foundation of efficient robot learning. Science 369(6506), 915\u2013916 (2020)","journal-title":"Science"},{"unstructured":"Micchelli, C.A., Pontil, M.: Kernels for multi\u2013task learning. In: NIPS, vol.\u00a086, p.\u00a089. Citeseer (2004)","key":"34_CR14"},{"unstructured":"Mukadam, M., Cheng, C.A., Fox, D., Boots, B., Ratliff, N.: Riemannian motion policy fusion through learnable lyapunov function reshaping. In: Conference on Robot Learning, pp. 204\u2013219. PMLR (2020)","key":"34_CR15"},{"unstructured":"Nachum, O., Gu, S., Lee, H., Levine, S.: Data-efficient hierarchical reinforcement learning. arXiv preprint arXiv:1805.08296 (2018)","key":"34_CR16"},{"doi-asserted-by":"crossref","unstructured":"Notomista, G., Mayya, S., Hutchinson, S., Egerstedt, M.: An optimal task allocation strategy for heterogeneous multi-robot systems. In: 2019 18th European Control Conference (ECC), pp. 2071\u20132076. IEEE (2019)","key":"34_CR17","DOI":"10.23919\/ECC.2019.8795895"},{"doi-asserted-by":"crossref","unstructured":"Notomista, G., Mayya, S., Selvaggio, M., Santos, M., Secchi, C.: A set-theoretic approach to multi-task execution and prioritization. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 9873\u20139879. IEEE (2020)","key":"34_CR18","DOI":"10.1109\/ICRA40945.2020.9196741"},{"unstructured":"Peng, X.B., Chang, M., Zhang, G., Abbeel, P., Levine, S.: MCP: learning composable hierarchical control with multiplicative compositional policies. arXiv preprint arXiv:1905.09808 (2019)","key":"34_CR19"},{"issue":"1","key":"34_CR20","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1111\/j.1934-6093.1999.tb00002.x","volume":"1","author":"JA Primbs","year":"1999","unstructured":"Primbs, J.A., Nevisti\u0107, V., Doyle, J.C.: Nonlinear optimal control: a control Lyapunov function and receding horizon perspective. Asian J. Control 1(1), 14\u201324 (1999)","journal-title":"Asian J. Control"},{"unstructured":"Qureshi, A.H., Johnson, J.J., Qin, Y., Henderson, T., Boots, B., Yip, M.C.: Composing task-agnostic policies with deep reinforcement learning. arXiv preprint arXiv:1905.10681 (2019)","key":"34_CR21"},{"unstructured":"Rana, M.A., et al.: Learning reactive motion policies in multiple task spaces from human demonstrations. In: Conference on Robot Learning, pp. 1457\u20131468. PMLR (2020)","key":"34_CR22"},{"unstructured":"Ratliff, N.D., Issac, J., Kappler, D., Birchfield, S., Fox, D.: Riemannian motion policies. arXiv preprint arXiv:1801.02854 (2018)","key":"34_CR23"},{"unstructured":"Ruder, S.: An overview of multi-task learning in deep neural networks. arXiv preprint arXiv:1706.05098 (2017)","key":"34_CR24"},{"unstructured":"Sahni, H., Kumar, S., Tejani, F., Isbell, C.: Learning to compose skills. arXiv preprint arXiv:1711.11289 (2017)","key":"34_CR25"},{"key":"34_CR26","first-page":"385","volume":"7","author":"A Schwartz","year":"1995","unstructured":"Schwartz, A., Thrun, S.: Finding structure in reinforcement learning. Adv. Neural. Inf. Process. Syst. 7, 385\u2013392 (1995)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"unstructured":"Sener, O., Koltun, V.: Multi-task learning as multi-objective optimization. arXiv preprint arXiv:1810.04650 (2018)","key":"34_CR27"},{"issue":"3","key":"34_CR28","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1007\/BF00992700","volume":"8","author":"SP Singh","year":"1992","unstructured":"Singh, S.P.: Transfer of learning by composing solutions of elemental sequential tasks. Mach. Learn. 8(3), 323\u2013339 (1992)","journal-title":"Mach. Learn."},{"unstructured":"Smith, V., Chiang, C.K., Sanjabi, M., Talwalkar, A.: Federated multi-task learning. arXiv preprint arXiv:1705.10467 (2017)","key":"34_CR29"},{"issue":"3","key":"34_CR30","doi-asserted-by":"publisher","first-page":"462","DOI":"10.1137\/0321028","volume":"21","author":"ED Sontag","year":"1983","unstructured":"Sontag, E.D.: A lyapunov-like characterization of asymptotic controllability. SIAM J. Control. Optim. 21(3), 462\u2013471 (1983)","journal-title":"SIAM J. Control. Optim."},{"issue":"2","key":"34_CR31","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1016\/0167-6911(89)90028-5","volume":"13","author":"ED Sontag","year":"1989","unstructured":"Sontag, E.D.: A \u201cuniversal\u2019\u2019 construction of Artstein\u2019s theorem on nonlinear stabilization. Syst. Control Lett. 13(2), 117\u2013123 (1989)","journal-title":"Syst. Control Lett."},{"unstructured":"Teh, Y.W., et al.: Distral: robust multitask reinforcement learning. arXiv preprint arXiv:1707.04175 (2017)","key":"34_CR32"},{"key":"34_CR33","first-page":"1856","volume":"22","author":"E Todorov","year":"2009","unstructured":"Todorov, E.: Compositionality of optimal control laws. Adv. Neural. Inf. Process. Syst. 22, 1856\u20131864 (2009)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"unstructured":"Van\u00a0Niekerk, B., James, S., Earle, A., Rosman, B.: Composing value functions in reinforcement learning. In: International Conference on Machine Learning, pp. 6401\u20136409. PMLR (2019)","key":"34_CR34"},{"key":"34_CR35","doi-asserted-by":"publisher","first-page":"5586","DOI":"10.1109\/TKDE.2021.3070203","volume":"34","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Yang, Q.: A survey on multi-task learning. IEEE Trans. Knowl. Data Eng. 34, 5586\u20135609 (2021)","journal-title":"IEEE Trans. Knowl. Data Eng."}],"container-title":["Springer Proceedings in Advanced Robotics","Distributed Autonomous Robotic Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-51497-5_34","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T16:55:09Z","timestamp":1731171309000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-51497-5_34"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031514968","9783031514975"],"references-count":35,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-51497-5_34","relation":{},"ISSN":["2511-1256","2511-1264"],"issn-type":[{"type":"print","value":"2511-1256"},{"type":"electronic","value":"2511-1264"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DARS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Symposium on Distributed Autonomous Robotic Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Montb\u00e9liard","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dars2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/dars2022.org\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}