论文标题
揭露挂毯:泛化和忘记的持续学习的相互作用
Unveiling the Tapestry: the Interplay of Generalization and Forgetting in Continual Learning
论文作者
论文摘要
在AI中,概括是指模型在与给定任务相关的分发数据上表现良好的能力,而不是对其进行训练的数据。对于AI代理才能表现出色,它还必须具有持续的学习能力,因此代理会逐步学习执行一系列任务,而不会忘记先前获得的知识来解决旧任务。直观地,任务中的概括使模型可以学习可以轻松应用于新颖任务的基本功能,从而促进了更快的学习并在持续学习框架内的后续任务中提高了性能。相反,持续的学习方法通常包括减轻灾难性遗忘的机制,以确保保留早期任务的知识。对任务的知识保存在增强目前正在进行的任务的概括方面发挥了作用。尽管两种能力的相互作用具有直观的吸引力,但现有的有关持续学习和概括的文献已经分别进行。在促进两个领域的研究的初步努力中,我们首先提供经验证据,表明这些领域中的每个领域都对另一个领域都有相互积极的影响。接下来,在这一发现的基础上,我们引入了一种简单有效的技术,称为形状文本一致性(STCR),该技术适应不断学习。 STCR学习每个任务的形状和纹理表示形式,从而增强概括,从而减轻遗忘。值得注意的是,广泛的实验验证了我们的STCR可以与现有的持续学习方法无缝集成,包括无重播方法。它的性能超过了这些连续的学习方法,或者与已建立的概括技术相结合时,其差距很大。
In AI, generalization refers to a model's ability to perform well on out-of-distribution data related to the given task, beyond the data it was trained on. For an AI agent to excel, it must also possess the continual learning capability, whereby an agent incrementally learns to perform a sequence of tasks without forgetting the previously acquired knowledge to solve the old tasks. Intuitively, generalization within a task allows the model to learn underlying features that can readily be applied to novel tasks, facilitating quicker learning and enhanced performance in subsequent tasks within a continual learning framework. Conversely, continual learning methods often include mechanisms to mitigate catastrophic forgetting, ensuring that knowledge from earlier tasks is retained. This preservation of knowledge over tasks plays a role in enhancing generalization for the ongoing task at hand. Despite the intuitive appeal of the interplay of both abilities, existing literature on continual learning and generalization has proceeded separately. In the preliminary effort to promote studies that bridge both fields, we first present empirical evidence showing that each of these fields has a mutually positive effect on the other. Next, building upon this finding, we introduce a simple and effective technique known as Shape-Texture Consistency Regularization (STCR), which caters to continual learning. STCR learns both shape and texture representations for each task, consequently enhancing generalization and thereby mitigating forgetting. Remarkably, extensive experiments validate that our STCR, can be seamlessly integrated with existing continual learning methods, including replay-free approaches. Its performance surpasses these continual learning methods in isolation or when combined with established generalization techniques by a large margin.