{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:27:21Z","timestamp":1750220841550,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":25,"publisher":"ACM","license":[{"start":{"date-parts":[[2019,11,5]],"date-time":"2019-11-05T00:00:00Z","timestamp":1572912000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CNS-1657350, CMMI-1831140"],"award-info":[{"award-number":["CNS-1657350, CMMI-1831140"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2019,11,5]]},"DOI":"10.1145\/3356471.3365229","type":"proceedings-article","created":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T12:18:47Z","timestamp":1572610727000},"page":"9-10","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Imitation Learning from Human-Generated Spatial-Temporal Data"],"prefix":"10.1145","author":[{"given":"Yanhua","family":"Li","sequence":"first","affiliation":[{"name":"Worcester Polytechnic Institute"}]},{"given":"Weixiao","family":"Huang","sequence":"additional","affiliation":[{"name":"Worcester Polytechnic Institute"}]}],"member":"320","published-online":{"date-parts":[[2019,11,5]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/1015330.1015430"},{"volume-title":"Psychology of learning and motivation.","author":"Berridge Kent C","key":"e_1_3_2_1_2_1","unstructured":"Kent C Berridge . 2000. Reward learning: Reinforcement, incentives, and expectations . In Psychology of learning and motivation. Vol. 40 . Elsevier , 223--278. Kent C Berridge. 2000. Reward learning: Reinforcement, incentives, and expectations. In Psychology of learning and motivation. Vol. 40. Elsevier, 223--278."},{"key":"e_1_3_2_1_3_1","volume-title":"Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics. 182--189","author":"Boularias Abdeslam","year":"2011","unstructured":"Abdeslam Boularias , Jens Kober , and Jan Peters . 2011 . Relative entropy inverse reinforcement learning . In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics. 182--189 . Abdeslam Boularias, Jens Kober, and Jan Peters. 2011. Relative entropy inverse reinforcement learning. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics. 182--189."},{"key":"e_1_3_2_1_4_1","unstructured":"DoorDash. 2019. DoorDash Services. https:\/\/www.doordash.com\/.  DoorDash. 2019. DoorDash Services. https:\/\/www.doordash.com\/."},{"key":"e_1_3_2_1_5_1","unstructured":"Chelsea Finn Sergey Levine and Pieter Abbeel. 2016. Guided cost learning: Deep inverse optimal control via policy optimization. In ICML.  Chelsea Finn Sergey Levine and Pieter Abbeel. 2016. Guided cost learning: Deep inverse optimal control via policy optimization. In ICML."},{"key":"e_1_3_2_1_6_1","unstructured":"Justin Fu Anoop Korattikara Sergey Levine and Sergio Guadarrama. 2019. From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following. In ICLR.  Justin Fu Anoop Korattikara Sergey Levine and Sergio Guadarrama. 2019. From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following. In ICLR."},{"key":"e_1_3_2_1_7_1","unstructured":"Jonathan Ho and Stefano Ermon. 2016. Generative adversarial imitation learning. In NeurIPS. 4565--4573.  Jonathan Ho and Stefano Ermon. 2016. Generative adversarial imitation learning. In NeurIPS. 4565--4573."},{"key":"e_1_3_2_1_8_1","volume-title":"Inverse Reinforcement Learning in Large State Spaces via Function Approximation. arXiv preprint arXiv:1707.09394","author":"Li Kun","year":"2017","unstructured":"Kun Li and Joel W Burdick . 2017. Inverse Reinforcement Learning in Large State Spaces via Function Approximation. arXiv preprint arXiv:1707.09394 ( 2017 ). Kun Li and Joel W Burdick. 2017. Inverse Reinforcement Learning in Large State Spaces via Function Approximation. arXiv preprint arXiv:1707.09394 (2017)."},{"key":"e_1_3_2_1_9_1","volume-title":"Meta inverse reinforcement learning via maximum reward sharing for human motion analysis. arXiv preprint arXiv:1710.03592","author":"Li Kun","year":"2017","unstructured":"Kun Li and Joel W Burdick . 2017. Meta inverse reinforcement learning via maximum reward sharing for human motion analysis. arXiv preprint arXiv:1710.03592 ( 2017 ). Kun Li and Joel W Burdick. 2017. Meta inverse reinforcement learning via maximum reward sharing for human motion analysis. arXiv preprint arXiv:1710.03592 (2017)."},{"key":"e_1_3_2_1_10_1","volume-title":"DHPA: Dynamic Human Preference Analytics Framework -- A Case Study on Taxi Drivers' Learning Curve Analysis. ACM Transactions on Intelligent Systems and Technology (TIST)","author":"Pan Menghai","year":"2019","unstructured":"Menghai Pan , Weixiao Huang , Yanhua Li , Xun Zhou , Zhenming Liu , Rui Song , Hui Lu , Zhihong Tian , and Jun Luo . 2019 . DHPA: Dynamic Human Preference Analytics Framework -- A Case Study on Taxi Drivers' Learning Curve Analysis. ACM Transactions on Intelligent Systems and Technology (TIST) (2019). Menghai Pan, Weixiao Huang, Yanhua Li, Xun Zhou, Zhenming Liu, Rui Song, Hui Lu, Zhihong Tian, and Jun Luo. 2019. DHPA: Dynamic Human Preference Analytics Framework -- A Case Study on Taxi Drivers' Learning Curve Analysis. ACM Transactions on Intelligent Systems and Technology (TIST) (2019)."},{"key":"e_1_3_2_1_11_1","unstructured":"Menghai Pan Yanhua Li Xun Zhou Zhenming Liu Rui Song Hui Lu and Jun Luo. 2019. Dissecting the Learning Curve of Taxi Drivers: A Data-Driven Approach. In SDM. SIAM.  Menghai Pan Yanhua Li Xun Zhou Zhenming Liu Rui Song Hui Lu and Jun Luo. 2019. Dissecting the Learning Curve of Taxi Drivers: A Data-Driven Approach. In SDM. SIAM."},{"key":"e_1_3_2_1_12_1","unstructured":"Postmates. 2019. Postmates Services. https:\/\/postmates.com\/.  Postmates. 2019. Postmates Services. https:\/\/postmates.com\/."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"Masamichi Shimosaka Junichi Sato Kazuhito Takenaka and Kentarou Hitomi. 2017. Fast inverse reinforcement learning with interval consistent graph for driving behavior prediction. In AAAI.  Masamichi Shimosaka Junichi Sato Kazuhito Takenaka and Kentarou Hitomi. 2017. Fast inverse reinforcement learning with interval consistent graph for driving behavior prediction. In AAAI.","DOI":"10.1609\/aaai.v31i1.10762"},{"key":"e_1_3_2_1_14_1","unstructured":"Umar Syed and Robert E Schapire. 2008. A game-theoretic approach to apprenticeship learning. In Advances in neural information processing systems. 1449--1456.  Umar Syed and Robert E Schapire. 2008. A game-theoretic approach to apprenticeship learning. In Advances in neural information processing systems. 1449--1456."},{"key":"e_1_3_2_1_15_1","unstructured":"TaskRabbit. 2019. TaskRabbit Services. https:\/\/www.taskrabbit.com\/.  TaskRabbit. 2019. TaskRabbit Services. https:\/\/www.taskrabbit.com\/."},{"key":"e_1_3_2_1_16_1","volume-title":"Behavioral cloning from observation. arXiv preprint arXiv:1805.01954","author":"Torabi Faraz","year":"2018","unstructured":"Faraz Torabi , Garrett Warnell , and Peter Stone . 2018. Behavioral cloning from observation. arXiv preprint arXiv:1805.01954 ( 2018 ). Faraz Torabi, Garrett Warnell, and Peter Stone. 2018. Behavioral cloning from observation. arXiv preprint arXiv:1805.01954 (2018)."},{"key":"e_1_3_2_1_17_1","volume-title":"12th international conference on travel behaviour research","author":"Wahba Mohamed","year":"2009","unstructured":"Mohamed Wahba and Amer Shalaby . 2009 . Learning-based departure time and path choice modelling for transit assignment under information provision: a theoretical framework . In 12th international conference on travel behaviour research , Jaipur, Rajasthan, India. Mohamed Wahba and Amer Shalaby. 2009. Learning-based departure time and path choice modelling for transit assignment under information provision: a theoretical framework. In 12th international conference on travel behaviour research, Jaipur, Rajasthan, India."},{"key":"e_1_3_2_1_18_1","volume":"201","author":"Waugh Kevin","unstructured":"Kevin Waugh , Brian D Ziebart , and J Andrew Bagnell. 201 3. Computational rationalization: The inverse equilibrium problem. arXiv preprint arXiv:1308.3506 (2013). Kevin Waugh, Brian D Ziebart, and J Andrew Bagnell. 2013. Computational rationalization: The inverse equilibrium problem. arXiv preprint arXiv:1308.3506 (2013).","journal-title":"J Andrew Bagnell."},{"key":"e_1_3_2_1_19_1","volume-title":"Data-Driven Inverse Learning of Passenger Preferences in Urban Public Transits. In CDC","author":"Wu Guojun","year":"2017","unstructured":"Guojun Wu , Yichen Ding , Yanhua Li , Jun Luo , Fan Zhang , and Jie Fu. [n. d.]. Data-Driven Inverse Learning of Passenger Preferences in Urban Public Transits. In CDC 2017 . IEEE. Guojun Wu, Yichen Ding, Yanhua Li, Jun Luo, Fan Zhang, and Jie Fu. [n. d.]. Data-Driven Inverse Learning of Passenger Preferences in Urban Public Transits. In CDC 2017. IEEE."},{"key":"e_1_3_2_1_20_1","unstructured":"Guojun Wu Yanhua Li Jie Bao Yu Zheng Jieping Ye and Jun Luo. 2018. Human-Centric Urban Transit Evaluation and Planning. In ICDM.  Guojun Wu Yanhua Li Jie Bao Yu Zheng Jieping Ye and Jun Luo. 2018. Human-Centric Urban Transit Evaluation and Planning. In ICDM."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CDC40024.2019.9029286"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2019.00194"},{"key":"e_1_3_2_1_23_1","volume-title":"The 10th International Conference on Autonomous Agents and Multiagent Systems-Volume 1. International Foundation for Autonomous Agents and Multiagent Systems, 207--214","author":"Ziebart Brian D","year":"2011","unstructured":"Brian D Ziebart , J Andrew Bagnell , and Anind K Dey . 2011 . Maximum causal entropy correlated equilibria for Markov games . In The 10th International Conference on Autonomous Agents and Multiagent Systems-Volume 1. International Foundation for Autonomous Agents and Multiagent Systems, 207--214 . Brian D Ziebart, J Andrew Bagnell, and Anind K Dey. 2011. Maximum causal entropy correlated equilibria for Markov games. In The 10th International Conference on Autonomous Agents and Multiagent Systems-Volume 1. International Foundation for Autonomous Agents and Multiagent Systems, 207--214."},{"key":"e_1_3_2_1_24_1","volume-title":"AAAI","volume":"8","author":"Ziebart Brian D","year":"2008","unstructured":"Brian D Ziebart , Andrew L Maas , J Andrew Bagnell , and Anind K Dey . 2008 . Maximum Entropy Inverse Reinforcement Learning .. In AAAI , Vol. 8 . Chicago, IL, USA, 1433--1438. Brian D Ziebart, Andrew L Maas, J Andrew Bagnell, and Anind K Dey. 2008. Maximum Entropy Inverse Reinforcement Learning.. In AAAI, Vol. 8. Chicago, IL, USA, 1433--1438."},{"key":"e_1_3_2_1_25_1","unstructured":"Brian D Ziebart Andrew L Maas J Andrew Bagnell and Anind K Dey. 2009. Human Behavior Modeling with Maximum Entropy Inverse Optimal Control. In AAAI. 92.  Brian D Ziebart Andrew L Maas J Andrew Bagnell and Anind K Dey. 2009. Human Behavior Modeling with Maximum Entropy Inverse Optimal Control. In AAAI. 92."}],"event":{"name":"SIGSPATIAL '19: 27th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems","sponsor":["SIGSPATIAL ACM Special Interest Group on Spatial Information"],"location":"Chicago IL USA","acronym":"SIGSPATIAL '19"},"container-title":["Proceedings of the 3rd ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3356471.3365229","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3356471.3365229","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3356471.3365229","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T23:22:54Z","timestamp":1750202574000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3356471.3365229"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,11,5]]},"references-count":25,"alternative-id":["10.1145\/3356471.3365229","10.1145\/3356471"],"URL":"https:\/\/doi.org\/10.1145\/3356471.3365229","relation":{},"subject":[],"published":{"date-parts":[[2019,11,5]]},"assertion":[{"value":"2019-11-05","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}