论文标题

自适应在线增量学习用于不断发展的数据流

Adaptive Online Incremental Learning for Evolving Data Streams

论文作者

Zhang, Si-si, Liu, Jian-wei, Zuo, Xin

论文摘要

近年来,人们对在线增量学习的兴趣不断增长。但是,该领域面临三个主要挑战。第一个主要困难是概念漂移,即,流数据到达时,流数据中的概率分布将发生变化。第二个主要困难是灾难性的遗忘,也就是说,忘记了我们在学习新知识之前学到的东西。我们经常忽略的最后一个是学习潜在表示。只有良好的潜在表示才能提高模型的预测准确性。我们的研究以这一观察结果为基础,并试图克服这些困难。为此,我们为不断发展的数据流(Aoil)提出了自适应在线增量学习。我们将自动编码器与内存模块一起使用,一方面,我们获得了输入的潜在特征,另一方面,根据使用内存模块的自动编码器的重建损失,我们可以成功地检测概念漂移的存在并触发更新机制,并及时调整模型参数。此外,我们将来自隐藏层的激活得出的特征分为两个部分,分别用于提取常见和私人特征。通过这种方法,该模型可以学习新即将到来的实例的私人功能,但不要忘记我们过去学到的东西(共享特征),这减少了灾难性遗忘的发生。同时,为了获得Fusion特征向量,我们使用自我注意力的机制有效地融合了提取的特征,从而进一步改善了潜在的表示。

Recent years have witnessed growing interests in online incremental learning. However, there are three major challenges in this area. The first major difficulty is concept drift, that is, the probability distribution in the streaming data would change as the data arrives. The second major difficulty is catastrophic forgetting, that is, forgetting what we have learned before when learning new knowledge. The last one we often ignore is the learning of the latent representation. Only good latent representation can improve the prediction accuracy of the model. Our research builds on this observation and attempts to overcome these difficulties. To this end, we propose an Adaptive Online Incremental Learning for evolving data streams (AOIL). We use auto-encoder with the memory module, on the one hand, we obtained the latent features of the input, on the other hand, according to the reconstruction loss of the auto-encoder with memory module, we could successfully detect the existence of concept drift and trigger the update mechanism, adjust the model parameters in time. In addition, we divide features, which are derived from the activation of the hidden layers, into two parts, which are used to extract the common and private features respectively. By means of this approach, the model could learn the private features of the new coming instances, but do not forget what we have learned in the past (shared features), which reduces the occurrence of catastrophic forgetting. At the same time, to get the fusion feature vector we use the self-attention mechanism to effectively fuse the extracted features, which further improved the latent representation learning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源