论文标题
研究生成LSTM网络中有效的学习和组成性
Investigating Efficient Learning and Compositionality in Generative LSTM Networks
论文作者
论文摘要
当将人类与人工智能进行比较时,一个主要区别是一个很明显的区别:人类可以从稀疏的数据集中概括,因为他们能够重新组合和重新集成构图的数据组件。为了调查高效学习的差异,约书亚·B·坦宁鲍姆(Joshua B. Tenenbaum)及其同事制定了角色挑战:首先,对算法进行了培训,以生成手写字符。在下一步中,提出了一种新类型的字符。有效的学习算法有望能够重新生成这个新角色,识别此字符的类似版本,生成其新变体,并创建全新的角色类型。过去,只有通过随机原始物提供的复杂算法来应对角色挑战。在这里,我们无需提供原语的问题就应对挑战。我们将使用一个进料层和一个LSTM层应用最小的复发性神经网络(RNN)模型,然后训练它以从一hot编码的输入中生成顺序的手写字符轨迹。为了管理未经训练的字符的重新生成,如果仅显示一个示例,我们引入了一个单弹性推理机制:仅将梯度信号反向传播到馈电层的权重,而LSTM层则未被触摸。我们表明,我们的模型能够通过重新组合先前学习的动态子结构来应对角色挑战,而该子结构在隐藏的LSTM状态中可见。以这种方式利用RNN的组成能力可能是弥合人工智能之间差距的重要一步。
When comparing human with artificial intelligence, one major difference is apparent: Humans can generalize very broadly from sparse data sets because they are able to recombine and reintegrate data components in compositional manners. To investigate differences in efficient learning, Joshua B. Tenenbaum and colleagues developed the character challenge: First an algorithm is trained in generating handwritten characters. In a next step, one version of a new type of character is presented. An efficient learning algorithm is expected to be able to re-generate this new character, to identify similar versions of this character, to generate new variants of it, and to create completely new character types. In the past, the character challenge was only met by complex algorithms that were provided with stochastic primitives. Here, we tackle the challenge without providing primitives. We apply a minimal recurrent neural network (RNN) model with one feedforward layer and one LSTM layer and train it to generate sequential handwritten character trajectories from one-hot encoded inputs. To manage the re-generation of untrained characters, when presented with only one example of them, we introduce a one-shot inference mechanism: the gradient signal is backpropagated to the feedforward layer weights only, leaving the LSTM layer untouched. We show that our model is able to meet the character challenge by recombining previously learned dynamic substructures, which are visible in the hidden LSTM states. Making use of the compositional abilities of RNNs in this way might be an important step towards bridging the gap between human and artificial intelligence.