论文标题
对信息流有关持续学习绩效的理论理解
Theoretical Understanding of the Information Flow on Continual Learning Performance
论文作者
论文摘要
持续学习(CL)是代理必须顺序从数据流中学习的设置。 CL绩效评估模型在保留以前的知识的同时,随着时间的推移,使用增量可用信息不断学习和解决新问题的能力。尽管以前有许多解决方案可以绕过学习过程中先前看到的任务的灾难性遗忘(CF),但其中大多数仍具有巨大的遗忘,昂贵的记忆成本,或者在学习新任务时对神经网络的行为缺乏理论上的理解。尽管在不同训练方面的CL性能降低的问题已经在经验上进行了广泛的研究,但从理论角度引起了足够的关注。在本文中,我们建立了一个概率框架,以分析通过网络中的各个层次的信息流以进行任务序列及其对学习绩效的影响。我们的目标是在学习新任务时优化层之间的信息保存,以管理整个层次的特定任务知识,同时在先前的任务上保持模型性能。特别是,我们研究CL性能与网络中信息流的关系,以回答“如何使用图层之间的信息流以减轻CF的知识?”。我们的分析提供了在渐进任务学习过程中层中信息适应的新颖见解。通过我们的实验,我们提供了经验证据,并实际上突出了多个任务之间的绩效提高。
Continual learning (CL) is a setting in which an agent has to learn from an incoming stream of data sequentially. CL performance evaluates the model's ability to continually learn and solve new problems with incremental available information over time while retaining previous knowledge. Despite the numerous previous solutions to bypass the catastrophic forgetting (CF) of previously seen tasks during the learning process, most of them still suffer significant forgetting, expensive memory cost, or lack of theoretical understanding of neural networks' conduct while learning new tasks. While the issue that CL performance degrades under different training regimes has been extensively studied empirically, insufficient attention has been paid from a theoretical angle. In this paper, we establish a probabilistic framework to analyze information flow through layers in networks for task sequences and its impact on learning performance. Our objective is to optimize the information preservation between layers while learning new tasks to manage task-specific knowledge passing throughout the layers while maintaining model performance on previous tasks. In particular, we study CL performance's relationship with information flow in the network to answer the question "How can knowledge of information flow between layers be used to alleviate CF?". Our analysis provides novel insights of information adaptation within the layers during the incremental task learning process. Through our experiments, we provide empirical evidence and practically highlight the performance improvement across multiple tasks.