论文标题
鲁棒的顺序神经过程
Robustifying Sequential Neural Processes
论文作者
论文摘要
当任务随着时间而变化时,元转移学习旨在通过元学习和转移学习来提高学习新任务的效率。尽管标准的关注在各种环境中都是有效的,但我们质疑其在改善元转移学习方面的有效性,因为学习的任务是动态的,并且上下文的数量可能会大大较小。在本文中,使用最近提出的元转移学习模型,顺序神经过程(SNP),我们首先从经验上表明,它遭受了在神经过程推断的功能中观察到的类似的不足问题。但是,我们进一步证明,与元学习环境不同,标准注意机制在元转移设置中无效。为了解决问题,我们提出了一种新的注意机制,经常性记忆重建(RMR),并证明提供了一个经常更新和通过相互作用重建的虚构环境对于实现元转移学习的有效关注至关重要。此外,将RMR纳入SNP,我们提出了细心的顺序神经过程-RMR(ASNP-RMR),并在各种任务中证明ASNP-RMR明显优于基准。
When tasks change over time, meta-transfer learning seeks to improve the efficiency of learning a new task via both meta-learning and transfer-learning. While the standard attention has been effective in a variety of settings, we question its effectiveness in improving meta-transfer learning since the tasks being learned are dynamic and the amount of context can be substantially smaller. In this paper, using a recently proposed meta-transfer learning model, Sequential Neural Processes (SNP), we first empirically show that it suffers from a similar underfitting problem observed in the functions inferred by Neural Processes. However, we further demonstrate that unlike the meta-learning setting, the standard attention mechanisms are not effective in meta-transfer setting. To resolve, we propose a new attention mechanism, Recurrent Memory Reconstruction (RMR), and demonstrate that providing an imaginary context that is recurrently updated and reconstructed with interaction is crucial in achieving effective attention for meta-transfer learning. Furthermore, incorporating RMR into SNP, we propose Attentive Sequential Neural Processes-RMR (ASNP-RMR) and demonstrate in various tasks that ASNP-RMR significantly outperforms the baselines.