{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,23]],"date-time":"2024-10-23T02:02:50Z","timestamp":1729648970053,"version":"3.28.0"},"reference-count":18,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2009,5]]},"DOI":"10.1109\/robot.2009.5152289","type":"proceedings-article","created":{"date-parts":[[2009,8,24]],"date-time":"2009-08-24T11:04:04Z","timestamp":1251111844000},"page":"2904-2909","source":"Crossref","is-referenced-by-count":2,"title":["Least absolute policy iteration for robust value function approximation"],"prefix":"10.1109","author":[{"given":"Masashi","family":"Sugiyama","sequence":"first","affiliation":[]},{"given":"Hirotaka","family":"Hachiya","sequence":"additional","affiliation":[]},{"given":"Hisashi","family":"Kashima","sequence":"additional","affiliation":[]},{"given":"Tetsuro","family":"Morimura","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.2307\/2975974"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1002\/rnc.670"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1162\/0899766053011528"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TNN.2007.899520"},{"key":"ref14","first-page":"759","article-title":"Eligibility traces for off-policy policy evaluation","author":"precup","year":"2000","journal-title":"Proc Seventh Int Conf Machine Learning"},{"journal-title":"Adaptive importance sampling for value function approximation in off-policy reinforcement learning","year":"2009","author":"hachiya","key":"ref15"},{"key":"ref16","first-page":"417","article-title":"Off-policy temporal-difference learning with function approximation","author":"precup","year":"2001","journal-title":"Proceedings of International Conference on Machine Learning"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511804441"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1023\/A:1012470815092"},{"key":"ref4","doi-asserted-by":"crossref","DOI":"10.1002\/0471725250","author":"huber","year":"1981","journal-title":"Robust Statistics"},{"key":"ref3","first-page":"1107","article-title":"Least-squares policy iteration","volume":"4","author":"lagoudakis","year":"2003","journal-title":"Journal of Machine Learning Research"},{"key":"ref6","first-page":"105","article-title":"Considering of risk in reinforcement learning","author":"herger","year":"1994","journal-title":"Proceedings of the 11th International Conference on Machine Learning"},{"key":"ref5","doi-asserted-by":"crossref","DOI":"10.1002\/0471725382","author":"rousseeuw","year":"1987","journal-title":"Robust Regression and Outlier Detection"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1023\/A:1017940631555"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/S0378-4266(02)00271-6"},{"journal-title":"Neuro-Dynamic Programming","year":"1996","key":"ref2"},{"journal-title":"Reinforcement Learning An Introduction","year":"1998","author":"sutton","key":"ref1"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1527\/tjsai.16.353"}],"event":{"name":"2009 IEEE International Conference on Robotics and Automation (ICRA)","start":{"date-parts":[[2009,5,12]]},"location":"Kobe","end":{"date-parts":[[2009,5,17]]}},"container-title":["2009 IEEE International Conference on Robotics and Automation"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx5\/5076472\/5152175\/05152289.pdf?arnumber=5152289","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2017,6,18]],"date-time":"2017-06-18T19:05:15Z","timestamp":1497812715000},"score":1,"resource":{"primary":{"URL":"http:\/\/ieeexplore.ieee.org\/document\/5152289\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2009,5]]},"references-count":18,"URL":"https:\/\/doi.org\/10.1109\/robot.2009.5152289","relation":{},"subject":[],"published":{"date-parts":[[2009,5]]}}}