{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T01:26:38Z","timestamp":1769045198455,"version":"3.49.0"},"reference-count":57,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2018,7,30]],"date-time":"2018-07-30T00:00:00Z","timestamp":1532908800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2018,8,31]]},"abstract":"<jats:p>Basketball is one of the world's most popular sports because of the agility and speed demonstrated by the players. This agility and speed makes designing controllers to realize robust control of basketball skills a challenge for physics-based character animation. The highly dynamic behaviors and precise manipulation of the ball that occur in the game are difficult to reproduce for simulated players. In this paper, we present an approach for learning robust basketball dribbling controllers from motion capture data. Our system decouples a basketball controller into locomotion control and arm control components and learns each component separately. To achieve robust control of the ball, we develop an efficient pipeline based on trajectory optimization and deep reinforcement learning and learn non-linear arm control policies. We also present a technique for learning skills and the transition between skills simultaneously. Our system is capable of learning robust controllers for various basketball dribbling skills, such as dribbling between the legs and crossover moves. The resulting control graphs enable a simulated player to perform transitions between these skills and respond to user interaction.<\/jats:p>","DOI":"10.1145\/3197517.3201315","type":"journal-article","created":{"date-parts":[[2018,7,31]],"date-time":"2018-07-31T15:56:23Z","timestamp":1533052583000},"page":"1-14","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":96,"title":["Learning basketball dribbling skills using trajectory optimization and deep reinforcement learning"],"prefix":"10.1145","volume":"37","author":[{"given":"Libin","family":"Liu","sequence":"first","affiliation":[{"name":"DeepMotion Inc."}]},{"given":"Jessica","family":"Hodgins","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University"}]}],"member":"320","published-online":{"date-parts":[[2018,7,30]]},"reference":[{"key":"e_1_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2012.325"},{"key":"e_1_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cag.2013.04.007"},{"key":"e_1_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/2366145.2366175"},{"key":"e_1_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.5555\/1703435.1703673"},{"key":"e_1_2_2_5_1","volume-title":"2010 IEEE\/RSJ International Conference on Intelligent Robots and Systems. 2890--2895","author":"B\u00e4tz Georg","unstructured":"Georg B\u00e4tz , Uwe Mettin , Alexander Schmidts , Michael Scheint , Dirk Wollherr , and Anton S. Shiriaev . 2010. Ball dribbling with an underactuated continuous-time control phase: Theory & experiments . In 2010 IEEE\/RSJ International Conference on Intelligent Robots and Systems. 2890--2895 . Georg B\u00e4tz, Uwe Mettin, Alexander Schmidts, Michael Scheint, Dirk Wollherr, and Anton S. Shiriaev. 2010. Ball dribbling with an underactuated continuous-time control phase: Theory & experiments. In 2010 IEEE\/RSJ International Conference on Intelligent Robots and Systems. 2890--2895."},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/1618452.1618516"},{"key":"e_1_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1781156"},{"key":"e_1_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2786784.2786802"},{"key":"e_1_2_2_9_1","unstructured":"Ga\u00ebl Guennebaud Beno\u00eft Jacob and others. 2010. Eigen v3. http:\/\/eigen.tuxfamily.org. (2010).  Ga\u00ebl Guennebaud Beno\u00eft Jacob and others. 2010. Eigen v3. http:\/\/eigen.tuxfamily.org. (2010)."},{"key":"e_1_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2366145.2366174"},{"key":"e_1_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.2011.6161022"},{"key":"e_1_2_2_12_1","volume-title":"Towards a New Evolutionary Computation. Studies in Fuzziness and Soft Computing","author":"Hansen Nikolaus","unstructured":"Nikolaus Hansen . 2006. The CMA Evolution Strategy: A Comparing Review . In Towards a New Evolutionary Computation. Studies in Fuzziness and Soft Computing , Vol. 192 . Springer Berlin Heidelberg , 75--102. Nikolaus Hansen. 2006. The CMA Evolution Strategy: A Comparing Review. In Towards a New Evolutionary Computation. Studies in Fuzziness and Soft Computing, Vol. 192. Springer Berlin Heidelberg, 75--102."},{"key":"e_1_2_2_13_1","volume-title":"Hausknecht and Peter Stone","author":"Matthew","year":"2015","unstructured":"Matthew J. Hausknecht and Peter Stone . 2015 . Deep Reinforcement Learning in Parameterized Action Space. CoRR abs\/1511.04143 (2015). http:\/\/arxiv.org\/abs\/1511.04143 Matthew J. Hausknecht and Peter Stone. 2015. Deep Reinforcement Learning in Parameterized Action Space. CoRR abs\/1511.04143 (2015). http:\/\/arxiv.org\/abs\/1511.04143"},{"key":"e_1_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/218380.218414"},{"key":"e_1_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/1599470.1599476"},{"key":"e_1_2_2_16_1","volume-title":"Kingma and Jimmy Ba","author":"Diederik","year":"2014","unstructured":"Diederik P. Kingma and Jimmy Ba . 2014 . Adam : A Method for Stochastic Optimization. CoRR abs\/1412.6980 (2014). http:\/\/arxiv.org\/abs\/1412.6980 Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. CoRR abs\/1412.6980 (2014). http:\/\/arxiv.org\/abs\/1412.6980"},{"key":"e_1_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/1141911.1141969"},{"key":"e_1_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2983616"},{"key":"e_1_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1781155"},{"key":"e_1_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/1882261.1866160"},{"key":"e_1_2_2_21_1","volume-title":"Proceedings of the 30th International Conference on Machine Learning","volume":"9","author":"Levine Sergey","year":"2013","unstructured":"Sergey Levine and Vladlen Koltun . 2013 . Guided Policy Search . In Proceedings of the 30th International Conference on Machine Learning , Vol. 28(3). 1-- 9 . Sergey Levine and Vladlen Koltun. 2013. Guided Policy Search. In Proceedings of the 30th International Conference on Machine Learning, Vol. 28(3). 1--9."},{"key":"e_1_2_2_22_1","volume-title":"Proceedings of the 31st International Conference on Machine Learning","volume":"837","author":"Levine Sergey","year":"2014","unstructured":"Sergey Levine and Vladlen Koltun . 2014 . Learning Complex Neural Network Policies with Trajectory Optimization . In Proceedings of the 31st International Conference on Machine Learning , Vol. 32(2). 829-- 837 . Sergey Levine and Vladlen Koltun. 2014. Learning Complex Neural Network Policies with Trajectory Optimization. In Proceedings of the 31st International Conference on Machine Learning, Vol. 32(2). 829--837."},{"key":"e_1_2_2_23_1","volume-title":"Continuous control with deep reinforcement learning. CoRR abs\/1509.02971","author":"Lillicrap Timothy P.","year":"2015","unstructured":"Timothy P. Lillicrap , Jonathan J. Hunt , Alexander Pritzel , Nicolas Heess , Tom Erez , Yuval Tassa , David Silver , and Daan Wierstra . 2015. Continuous control with deep reinforcement learning. CoRR abs\/1509.02971 ( 2015 ). http:\/\/arxiv.org\/abs\/1509.02971 Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. 2015. Continuous control with deep reinforcement learning. CoRR abs\/1509.02971 (2015). http:\/\/arxiv.org\/abs\/1509.02971"},{"key":"e_1_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531365"},{"key":"e_1_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.5555\/1218064.1218093"},{"key":"e_1_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3083723"},{"key":"e_1_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/2893476"},{"key":"e_1_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/2508363.2508427"},{"key":"e_1_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531386"},{"key":"e_1_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/1276377.1276385"},{"key":"e_1_2_2_31_1","volume-title":"Human-level control through deep reinforcement learning. Nature 518, 7540 (26","author":"Mnih Volodymyr","year":"2015","unstructured":"Volodymyr Mnih , Koray Kavukcuoglu , David Silver , Andrei A. Rusu , Joel Veness , Marc G. Bellemare , Alex Graves , Martin Riedmiller , Andreas K. Fidjeland , Georg Ostrovski , Stig Petersen , Charles Beattie , Amir Sadik , Ioannis Antonoglou , Helen King , Dharshan Kumaran , Daan Wierstra , Shane Legg , and Demis Hassabis . 2015b. Human-level control through deep reinforcement learning. Nature 518, 7540 (26 Feb 2015 ), 529--533. Letter . Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. 2015b. Human-level control through deep reinforcement learning. Nature 518, 7540 (26 Feb 2015), 529--533. Letter."},{"key":"e_1_2_2_32_1","doi-asserted-by":"crossref","unstructured":"Volodymyr Mnih Koray Kavukcuoglu David Silver Andrei A. Rusu Joel Veness and etal 2015a. Human-level control through deep reinforcement learning. Nature 518 7540 (26 Feb 2015) 529--533. Letter.  Volodymyr Mnih Koray Kavukcuoglu David Silver Andrei A. Rusu Joel Veness and et al. 2015a. Human-level control through deep reinforcement learning. Nature 518 7540 (26 Feb 2015) 529--533. Letter.","DOI":"10.1038\/nature14236"},{"key":"e_1_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1778808"},{"key":"e_1_2_2_34_1","volume-title":"Todorov","author":"Mordatch Igor","year":"2015","unstructured":"Igor Mordatch , Kendall Lowrey , Galen Andrew , Zoran Popovic , and Emanuel V . Todorov . 2015 . Interactive Control of Diverse Complex Characters with Neural Networks. In Advances in Neural Information Processing Systems 28. 3114--3122. Igor Mordatch, Kendall Lowrey, Galen Andrew, Zoran Popovic, and Emanuel V. Todorov. 2015. Interactive Control of Diverse Complex Characters with Neural Networks. In Advances in Neural Information Processing Systems 28. 3114--3122."},{"key":"e_1_2_2_35_1","volume-title":"Proceedings of the ACM SIGGRAPH\/Eurographics Symposium on Computer Animation (SCA '12)","author":"Mordatch Igor","year":"2012","unstructured":"Igor Mordatch , Zoran Popovi\u0107 , and Emanuel Todorov . 2012 a. Contact-invariant Optimization for Hand Manipulation . In Proceedings of the ACM SIGGRAPH\/Eurographics Symposium on Computer Animation (SCA '12) . 137--144. Igor Mordatch, Zoran Popovi\u0107, and Emanuel Todorov. 2012a. Contact-invariant Optimization for Hand Manipulation. In Proceedings of the ACM SIGGRAPH\/Eurographics Symposium on Computer Animation (SCA '12). 137--144."},{"key":"e_1_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/2185520.2185539"},{"key":"e_1_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531387"},{"key":"e_1_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073602"},{"key":"e_1_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3099564.3099567"},{"key":"e_1_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/1073368.1073413"},{"key":"e_1_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/TRO.2012.2205493"},{"key":"e_1_2_2_42_1","volume-title":"Evolution Strategies as a Scalable Alternative to Reinforcement Learning. ArXiv e-prints (March","author":"Salimans Tim","year":"2017","unstructured":"Tim Salimans , Jonathan Ho , Xi Chen , Szymon Sidor , and Ilya Sutskever . 2017. Evolution Strategies as a Scalable Alternative to Reinforcement Learning. ArXiv e-prints (March 2017 ). arXiv:stat.ML\/1703.03864 Tim Salimans, Jonathan Ho, Xi Chen, Szymon Sidor, and Ilya Sutskever. 2017. Evolution Strategies as a Scalable Alternative to Reinforcement Learning. ArXiv e-prints (March 2017). arXiv:stat.ML\/1703.03864"},{"key":"e_1_2_2_43_1","volume-title":"Trust Region Policy Optimization. In The 32nd International Conference on Machine Learning. 1889--1897","author":"Schulman John","year":"2015","unstructured":"John Schulman , Sergey Levine , Pieter Abbeel , Michael Jordan , and Philipp Moritz . 2015 a. Trust Region Policy Optimization. In The 32nd International Conference on Machine Learning. 1889--1897 . John Schulman, Sergey Levine, Pieter Abbeel, Michael Jordan, and Philipp Moritz. 2015a. Trust Region Policy Optimization. In The 32nd International Conference on Machine Learning. 1889--1897."},{"key":"e_1_2_2_44_1","volume-title":"High-Dimensional Continuous Control Using Generalized Advantage Estimation. CoRR abs\/1506.02438","author":"Schulman John","year":"2015","unstructured":"John Schulman , Philipp Moritz , Sergey Levine , Michael I. Jordan , and Pieter Abbeel . 2015b. High-Dimensional Continuous Control Using Generalized Advantage Estimation. CoRR abs\/1506.02438 ( 2015 ). http:\/\/arxiv.org\/abs\/1506.02438 John Schulman, Philipp Moritz, Sergey Levine, Michael I. Jordan, and Pieter Abbeel. 2015b. High-Dimensional Continuous Control Using Generalized Advantage Estimation. CoRR abs\/1506.02438 (2015). http:\/\/arxiv.org\/abs\/1506.02438"},{"key":"e_1_2_2_45_1","volume-title":"Deterministic Policy Gradient Algorithms. In The 31st International Conference on Machine Learning. 387--395","author":"Silver David","year":"2014","unstructured":"David Silver , Guy Lever , Nicolas Heess , Thomas Degris , Daan Wierstra , and Martin Riedmiller . 2014 . Deterministic Policy Gradient Algorithms. In The 31st International Conference on Machine Learning. 387--395 . David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin Riedmiller. 2014. Deterministic Policy Gradient Algorithms. In The 31st International Conference on Machine Learning. 387--395."},{"key":"e_1_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/2601097.2601121"},{"key":"e_1_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/2010324.1964953"},{"key":"e_1_2_2_48_1","volume-title":"Theano: A Python framework for fast computation of mathematical expressions. arXiv e-prints abs\/1605.02688 (May","author":"Team Theano Development","year":"2016","unstructured":"Theano Development Team . 2016 . Theano: A Python framework for fast computation of mathematical expressions. arXiv e-prints abs\/1605.02688 (May 2016). http:\/\/arxiv.org\/abs\/1605.02688 Theano Development Team. 2016. Theano: A Python framework for fast computation of mathematical expressions. arXiv e-prints abs\/1605.02688 (May 2016). http:\/\/arxiv.org\/abs\/1605.02688"},{"key":"e_1_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/1276377.1276386"},{"key":"e_1_2_2_50_1","volume-title":"Reinforcement Learning in Continuous State and Action Spaces","author":"van Hasselt Hado","unstructured":"Hado van Hasselt . 2012. Reinforcement Learning in Continuous State and Action Spaces . Springer , Berlin, Heidelberg , 207--251. Hado van Hasselt. 2012. Reinforcement Learning in Continuous State and Action Spaces. Springer, Berlin, Heidelberg, 207--251."},{"key":"e_1_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531366"},{"key":"e_1_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1778810"},{"key":"e_1_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.12595"},{"key":"e_1_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/2185520.2185537"},{"key":"e_1_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/1276377.1276509"},{"key":"e_1_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/2508363.2508412"},{"key":"e_1_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2014.2330610"}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3197517.3201315","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3197517.3201315","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T01:39:44Z","timestamp":1750210784000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3197517.3201315"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,7,30]]},"references-count":57,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2018,8,31]]}},"alternative-id":["10.1145\/3197517.3201315"],"URL":"https:\/\/doi.org\/10.1145\/3197517.3201315","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2018,7,30]]},"assertion":[{"value":"2018-07-30","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}