{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T16:25:32Z","timestamp":1772555132736,"version":"3.50.1"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,9,24]],"date-time":"2023-09-24T00:00:00Z","timestamp":1695513600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,9,24]],"date-time":"2023-09-24T00:00:00Z","timestamp":1695513600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,9,24]]},"DOI":"10.1109\/itsc57777.2023.10422310","type":"proceedings-article","created":{"date-parts":[[2024,2,13]],"date-time":"2024-02-13T23:32:39Z","timestamp":1707867159000},"page":"4750-4755","source":"Crossref","is-referenced-by-count":2,"title":["Enhancing Joint Behavior Modeling with Route Choice Using Adversarial Inverse Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Daichi","family":"Ogawa","sequence":"first","affiliation":[{"name":"University of Tokyo,Department of Civil Engineering,Tokyo,Japan"}]},{"given":"Eiji","family":"Hato","sequence":"additional","affiliation":[{"name":"University of Tokyo,Department of Civil Engineering,Tokyo,Japan"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.trb.2013.07.012"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1177\/0278364920961809"},{"key":"ref3","article-title":"Learning robust rewards with adversarial inverse reinforcement learning","author":"Fu","year":"2017","journal-title":"arXiv preprint"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/S1389-0417(01)00015-8"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.trc.2017.10.001"},{"key":"ref6","article-title":"Generative adversarial nets","author":"Courville","year":"2014","journal-title":"Advanc in Neural"},{"key":"ref7","first-page":"29","article-title":"Generative adversarial imitation learning","author":"Ho","year":"2016","journal-title":"Advances in neural information processing systems"},{"key":"ref8","article-title":"Algorithms for inverse reinforcement learning","author":"Ng","year":"2000","journal-title":"Icml"},{"key":"ref9","article-title":"A connection between generative adversarial networks, inverse reinforcement learning, and energy-based models","author":"Finn","year":"2016","journal-title":"arXiv preprint"},{"key":"ref10","first-page":"278","article-title":"Policy invariance under reward transformations: Theory and application to reward shaping","author":"Ng","year":"1999","journal-title":"Icml"},{"key":"ref11","first-page":"793","article-title":"Deep counterfactual regret minimization","volume-title":"International conference on machine learning","author":"Brown","year":"2019"},{"key":"ref12","first-page":"15","article-title":"Reinforcement learning to play an optimal Nash equilibrium in team Markov games","author":"Wang","year":"2002","journal-title":"Advances in neural information processing systems"},{"key":"ref13","first-page":"31","article-title":"Multi-agent generative adversarial imitation learning","author":"Song","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ISIT.2004.1365067"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/bth100"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1214\/aoms\/1177730090"},{"key":"ref17","author":"Minami","year":"1994","journal-title":"Variance estimation for simultaneous response growth curve models"},{"key":"ref18","first-page":"7194","article-title":"Multi-agent adversarial inverse reinforcement learning","volume-title":"International Conference on Machine Learning","author":"Yu","year":"2019"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1111\/1468-0262.00429"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1016\/j.jet.2005.09.005"},{"key":"ref21","volume-title":"Planet dump","year":"2017"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.trc.2023.104079"},{"issue":"3","key":"ref23","first-page":"261","article-title":"SciPy 1.0 Contributors","volume":"17","author":"Pauli","year":"2020","journal-title":"SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods"},{"key":"ref24","article-title":"Proximal policy optimization algorithms","author":"Schulman","year":"2017","journal-title":"arXiv preprint"},{"key":"ref25","first-page":"8024","article-title":"PyTorch: An Imperative Style, High-Performance Deep Learning Library","volume-title":"Advances in Neural Information Processing Systems","volume":"32","author":"Paszke","year":"2019"}],"event":{"name":"2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC)","location":"Bilbao, Spain","start":{"date-parts":[[2023,9,24]]},"end":{"date-parts":[[2023,9,28]]}},"container-title":["2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10420842\/10420843\/10422310.pdf?arnumber=10422310","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T20:40:36Z","timestamp":1710362436000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10422310\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,24]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/itsc57777.2023.10422310","relation":{},"subject":[],"published":{"date-parts":[[2023,9,24]]}}}