{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T04:38:34Z","timestamp":1750307914264,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":11,"publisher":"ACM","license":[{"start":{"date-parts":[[2007,3,10]],"date-time":"2007-03-10T00:00:00Z","timestamp":1173484800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2007,3,10]]},"DOI":"10.1145\/1228716.1228726","type":"proceedings-article","created":{"date-parts":[[2007,4,5]],"date-time":"2007-04-05T19:41:00Z","timestamp":1175802060000},"page":"65-72","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":31,"title":["Efficient model learning for dialog management"],"prefix":"10.1145","author":[{"given":"Finale","family":"Doshi","sequence":"first","affiliation":[{"name":"CSAIL MIT, Cambridge, MA"}]},{"given":"Nicholas","family":"Roy","sequence":"additional","affiliation":[{"name":"CSAIL MIT, Cambridge, MA"}]}],"member":"320","published-online":{"date-parts":[[2007,3,10]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"150","volume-title":"Model based bayesian exploration","author":"Dearden R.","year":"1999","unstructured":"R. Dearden , N. Friedman , and D. Andre . Model based bayesian exploration . pages 150 -- 159 , 1999 . R. Dearden, N. Friedman, and D. Andre. Model based bayesian exploration. pages 150--159, 1999."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.5555\/3091622.3091655"},{"key":"e_1_3_2_1_3_1","volume-title":"Workshop on Non-Stationarity in Reinforcement Learning at the ECML","author":"Jaulmes R.","year":"2005","unstructured":"R. Jaulmes , J. Pineau , and D. Precup . Learning in non-stationary partially observable markov decision processes . Workshop on Non-Stationarity in Reinforcement Learning at the ECML , 2005 . R. Jaulmes, J. Pineau, and D. Precup. Learning in non-stationary partially observable markov decision processes. Workshop on Non-Stationarity in Reinforcement Learning at the ECML, 2005."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.3115\/1117562.1117566"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.5555\/1088001"},{"key":"e_1_3_2_1_6_1","volume-title":"Point-based value iteration: An anytime algorithm for pomdps","author":"Pineau J.","year":"2003","unstructured":"J. Pineau , G. Gordon , and S. Thrun . Point-based value iteration: An anytime algorithm for pomdps , 2003 . J. Pineau, G. Gordon, and S. Thrun. Point-based value iteration: An anytime algorithm for pomdps, 2003."},{"key":"e_1_3_2_1_7_1","volume-title":"Workshop on Hierarchy and Memory in Reinforcement Learning (ICML)","author":"Pineau J.","year":"2001","unstructured":"J. Pineau , N. Roy , and S. Thrun . A hierarchical approach to pomdp planning and execution . In Workshop on Hierarchy and Memory in Reinforcement Learning (ICML) , June 2001 . J. Pineau, N. Roy, and S. Thrun. A hierarchical approach to pomdp planning and execution. In Workshop on Hierarchy and Memory in Reinforcement Learning (ICML), June 2001."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.18626"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.3115\/1075218.1075231"},{"key":"e_1_3_2_1_11_1","volume-title":"Proceedings of the IEEE ASRU Workshop","author":"Williams J.","year":"2005","unstructured":"J. Williams and S. Young . Scaling up pomdps for dialogue management: The lhsummary pomdpla method . In Proceedings of the IEEE ASRU Workshop , 2005 . J. Williams and S. Young. Scaling up pomdps for dialogue management: The lhsummary pomdpla method. In Proceedings of the IEEE ASRU Workshop, 2005."},{"key":"e_1_3_2_1_12_1","volume-title":"Proceedings of SIGdial Workshop on Discourse and Dialogue 2005","author":"Williams J. D.","year":"2005","unstructured":"J. D. Williams , P. Poupart , and S. Young . Partially observable markov decision processes with continuous observations for dialogue management . In Proceedings of SIGdial Workshop on Discourse and Dialogue 2005 , 2005 . J. D. Williams, P. Poupart, and S. Young. Partially observable markov decision processes with continuous observations for dialogue management. In Proceedings of SIGdial Workshop on Discourse and Dialogue 2005, 2005."}],"event":{"name":"HRI07: International Conference on Human Robot Interaction","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence","ACM Association for Computing Machinery","SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"Arlington Virginia USA","acronym":"HRI07"},"container-title":["Proceedings of the ACM\/IEEE international conference on Human-robot interaction"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/1228716.1228726","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/1228716.1228726","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T14:47:56Z","timestamp":1750258076000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/1228716.1228726"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2007,3,10]]},"references-count":11,"alternative-id":["10.1145\/1228716.1228726","10.1145\/1228716"],"URL":"https:\/\/doi.org\/10.1145\/1228716.1228726","relation":{},"subject":[],"published":{"date-parts":[[2007,3,10]]},"assertion":[{"value":"2007-03-10","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}