{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T06:45:27Z","timestamp":1771051527819,"version":"3.50.1"},"publisher-location":"California","reference-count":0,"publisher":"International Joint Conferences on Artificial Intelligence Organization","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,8]]},"abstract":"<jats:p>Despite significant progress, deep reinforcement learning (RL) suffers from data-inefficiency and limited generalization. Recent efforts apply meta-learning to learn a meta-learner from a set of RL tasks such that a novel but related task could be solved quickly. Though specific in some ways, different tasks in meta-RL are generally similar at a high level. However, most meta-RL methods do not explicitly and adequately model the specific and shared information among different tasks, which limits their ability to learn training tasks and to generalize to novel tasks. In this paper, we propose to capture the shared information on the one hand and meta-learn how to quickly abstract the specific information about a task on the other hand. Methodologically, we train an SGD meta-learner to quickly optimize a task encoder for each task, which generates a task embedding based on past experience. Meanwhile, we learn a policy which is shared across all tasks and conditioned on task embeddings. Empirical results on four simulated tasks demonstrate that our method has better learning capacity on both training and novel tasks and attains up to 3 to 4 times higher returns compared to baselines.<\/jats:p>","DOI":"10.24963\/ijcai.2019\/387","type":"proceedings-article","created":{"date-parts":[[2019,7,28]],"date-time":"2019-07-28T07:46:05Z","timestamp":1564299965000},"page":"2794-2800","source":"Crossref","is-referenced-by-count":19,"title":["Meta Reinforcement Learning with Task Embedding and Shared Policy"],"prefix":"10.24963","author":[{"given":"Lin","family":"Lan","sequence":"first","affiliation":[{"name":"MOE NSKEY Lab, Xi\u2019an Jiaotong University, China"}]},{"given":"Zhenguo","family":"Li","sequence":"additional","affiliation":[{"name":"Huawei Noah's Ark Lab"}]},{"given":"Xiaohong","family":"Guan","sequence":"additional","affiliation":[{"name":"MOE NSKEY Lab, Xi\u2019an Jiaotong University, China"},{"name":"Shenzhen Research School, Xi\u2019an Jiaotong University, China"},{"name":"Department of Automation and NLIST Lab, Tsinghua University, China"}]},{"given":"Pinghui","family":"Wang","sequence":"additional","affiliation":[{"name":"Shenzhen Research School, Xi\u2019an Jiaotong University, China"},{"name":"MOE NSKEY Lab, Xi\u2019an Jiaotong University, China"}]}],"member":"10584","event":{"name":"Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}","theme":"Artificial Intelligence","location":"Macao, China","acronym":"IJCAI-2019","number":"28","sponsor":["International Joint Conferences on Artificial Intelligence Organization (IJCAI)"],"start":{"date-parts":[[2019,8,10]]},"end":{"date-parts":[[2019,8,16]]}},"container-title":["Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence"],"original-title":[],"deposited":{"date-parts":[[2019,7,28]],"date-time":"2019-07-28T07:48:57Z","timestamp":1564300137000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.ijcai.org\/proceedings\/2019\/387"}},"subtitle":[],"proceedings-subject":"Artificial Intelligence Research Articles","short-title":[],"issued":{"date-parts":[[2019,8]]},"references-count":0,"URL":"https:\/\/doi.org\/10.24963\/ijcai.2019\/387","relation":{},"subject":[],"published":{"date-parts":[[2019,8]]}}}