{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2022,3,30]],"date-time":"2022-03-30T02:24:48Z","timestamp":1648607088044},"reference-count":0,"publisher":"International Information and Engineering Technology Association","issue":"2-3","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Revue d'intelligence artificielle"],"published-print":{"date-parts":[[2006,6,1]]},"DOI":"10.3166\/ria.20.275-310","type":"journal-article","created":{"date-parts":[[2008,8,19]],"date-time":"2008-08-19T07:32:40Z","timestamp":1219131160000},"page":"275-310","source":"Crossref","is-referenced-by-count":0,"title":["Apprentissage par renforcement dans le cadre des processus d\u00e9cisionnels de Markov factoris\u00e9s observables dans le d\u00e9sordre. Etude exp\u00e9rimentale du Q-Learning parall\u00e8le appliqu\u00e9 aux probl\u00e8mes du labyrinthe et du New York Driving"],"prefix":"10.18280","volume":"20","author":[{"given":"Guillaume J.","family":"Laurent","sequence":"first","affiliation":[]},{"given":"Emmanuel","family":"Piat","sequence":"additional","affiliation":[]}],"member":"7618","container-title":["Revue d'intelligence artificielle"],"original-title":[],"deposited":{"date-parts":[[2019,7,16]],"date-time":"2019-07-16T15:34:56Z","timestamp":1563291296000},"score":1,"resource":{"primary":{"URL":"http:\/\/ria.revuesonline.com\/article.jsp?articleId=8137"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2006,6,1]]},"references-count":0,"journal-issue":{"issue":"2-3","published-print":{"date-parts":[[2006,6,1]]}},"URL":"https:\/\/doi.org\/10.3166\/ria.20.275-310","relation":{},"ISSN":["0992-499X"],"issn-type":[{"value":"0992-499X","type":"print"}],"subject":[],"published":{"date-parts":[[2006,6,1]]}}}