{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T00:11:51Z","timestamp":1769127111694,"version":"3.49.0"},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T00:00:00Z","timestamp":1764633600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T00:00:00Z","timestamp":1764633600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,2]]},"DOI":"10.1109\/icar65334.2025.11338705","type":"proceedings-article","created":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T21:06:47Z","timestamp":1769029607000},"page":"214-219","source":"Crossref","is-referenced-by-count":0,"title":["Adiabatic Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Pavel","family":"Osinenko","sequence":"first","affiliation":[{"name":"Center for Digital Engineering,Moscow,Russia"}]},{"given":"Grigory","family":"Yaremenko","sequence":"additional","affiliation":[{"name":"Center for Digital Engineering,Moscow,Russia"}]},{"given":"Danil","family":"Belov","sequence":"additional","affiliation":[{"name":"Center for Digital Engineering,Moscow,Russia"}]},{"given":"Alexander","family":"Gepperth","sequence":"additional","affiliation":[{"name":"Fulda University of Applied Sciences,Fulda,Germany"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2020.11.050"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN55064.2022.9892384"},{"key":"ref3","article-title":"Beyond supervised continual learning: a review","author":"Bagus","year":"2022"},{"key":"ref4","article-title":"Model-free generative replay for lifelong reinforcement learning: Application to starcraft-2","author":"Daniels","year":"2022"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN55064.2022.9892212"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-021-10599-3"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1.13673"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN60899.2024.10651381"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW63382.2024.00423"},{"key":"ref10","article-title":"Understanding continual learning settings with data distribution drift analysis","author":"Lesort","year":"2021"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2019.12.004"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1016\/s0079-7421(08)60536-8"},{"key":"ref13","article-title":"Playing atari with deep reinforcement learning","author":"Mnih","year":"2013"},{"key":"ref14","first-page":"1561","article-title":"Understanding and mitigating the limitations of prioritized experience replay","author":"Pan","year":"2022","journal-title":"Uncertainty in Artificial Intelligence"},{"key":"ref15","article-title":"A comprehensive, application-oriented study of catastrophic forgetting in dnns","volume-title":"ICLR International Conference on Learning Representations","author":"Pf\u00fclb"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1049\/cvi2.70013"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1037\/0033-295x.97.2.285"},{"key":"ref18","article-title":"Prioritized experience replay","author":"Schaul","year":"2015"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/s10846-022-01603-6"},{"key":"ref20","article-title":"Coom: a game benchmark for continual reinforcement learning","volume":"36","author":"Tomilin","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref21","article-title":"Discorl: Continual reinforcement learning via policy distillation","author":"Traor\u00e9","year":"2019"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-022-00568-3"},{"key":"ref23","article-title":"Continual learning: Applications and the road forward","author":"Verwimp","year":"2023","journal-title":"Transactions on Machine Learning Research (TMLR), 2024"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3147.3165"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2024.3367329\/mm1"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2024.3498346\/mm1"},{"key":"ref27","first-page":"28496","article-title":"Continual world: A robotic benchmark for continual reinforcement learning","volume":"34","author":"Wo\u0142czyk","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref28","first-page":"6304","article-title":"Disentangling transfer in continual reinforcement learning","volume-title":"Advances in Neural Information Processing Systems","volume":"35","author":"Wolczyk","year":"2022"}],"event":{"name":"2025 IEEE International Conference on Advanced Robotics (ICAR)","location":"San Juan, Argentina","start":{"date-parts":[[2025,12,2]]},"end":{"date-parts":[[2025,12,5]]}},"container-title":["2025 IEEE International Conference on Advanced Robotics (ICAR)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11338521\/11338600\/11338705.pdf?arnumber=11338705","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T08:01:54Z","timestamp":1769068914000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11338705\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,2]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/icar65334.2025.11338705","relation":{},"subject":[],"published":{"date-parts":[[2025,12,2]]}}}