论文标题
走向持续的加强学习:评论和观点
Towards Continual Reinforcement Learning: A Review and Perspectives
论文作者
论文摘要
在本文中,我们旨在提供有关持续强化学习(RL)(也称为终身或非平稳RL)的不同表述和方法的文献综述。首先,我们讨论RL为什么自然地研究持续学习的观点。然后,我们通过数学表征非平稳性的两个关键属性,即范围和驱动程序非平稳性来提供不同连续RL公式的分类学。这提供了各种配方的统一视图。接下来,我们审查并提出了连续RL方法的分类法。我们继续讨论对持续RL代理的评估,并提供了文献中使用的基准和重要指标,以了解代理性能。最后,我们重点介绍了弥合持续RL的当前状态与神经科学发现之间的差距的挑战。虽然仍在早期,但对RL的研究有望发展出更好的增量增强学习者,这些学习者可以在非平稳性起着至关重要的作用的日益现实的应用中发挥作用。这些包括诸如医疗保健,教育,物流和机器人技术领域的应用程序。
In this article, we aim to provide a literature review of different formulations and approaches to continual reinforcement learning (RL), also known as lifelong or non-stationary RL. We begin by discussing our perspective on why RL is a natural fit for studying continual learning. We then provide a taxonomy of different continual RL formulations by mathematically characterizing two key properties of non-stationarity, namely, the scope and driver non-stationarity. This offers a unified view of various formulations. Next, we review and present a taxonomy of continual RL approaches. We go on to discuss evaluation of continual RL agents, providing an overview of benchmarks used in the literature and important metrics for understanding agent performance. Finally, we highlight open problems and challenges in bridging the gap between the current state of continual RL and findings in neuroscience. While still in its early days, the study of continual RL has the promise to develop better incremental reinforcement learners that can function in increasingly realistic applications where non-stationarity plays a vital role. These include applications such as those in the fields of healthcare, education, logistics, and robotics.