论文标题
神经:神经解码的封闭形式的逆增强学习
NeuRL: Closed-form Inverse Reinforcement Learning for Neural Decoding
论文作者
论文摘要
当前的神经解码方法通常旨在通过监督学习来基于神经活动来解释行为。但是,由于通常在学科学习与他们对长期奖励的期望之间存在牢固的联系,因此我们提出了一种神经,这是一种逆增强学习方法,((1)从受试者的收集轨迹中提取固有的奖励功能,((2)将神经信号映射到这种内在的奖励,以对这种内在的奖励,以说明在行为和(3)行为中的长期依赖,以提出符合(3)的行为。基于这些看不见的神经信号的固有奖励值的相应鲍尔茨曼政策。我们表明,与监督方法相比,神经l可以提高更好的概括和改善的解码性能。我们研究大鼠在响应准备任务中的行为,并评估模拟抑制和每次试验行为预测中神经的性能。通过将明确的功能角色分配给定义的神经元种群,我们的方法为具有可测试预测的复杂神经元数据提供了新的解释工具。在每个试验行为预测中,与传统方法相比,我们的方法进一步提高了高达15%的精度。
Current neural decoding methods typically aim at explaining behavior based on neural activity via supervised learning. However, since generally there is a strong connection between learning of subjects and their expectations on long-term rewards, we propose NeuRL, an inverse reinforcement learning approach that (1) extracts an intrinsic reward function from collected trajectories of a subject in closed form, (2) maps neural signals to this intrinsic reward to account for long-term dependencies in the behavior and (3) predicts the simulated behavior for unseen neural signals by extracting Q-values and the corresponding Boltzmann policy based on the intrinsic reward values for these unseen neural signals. We show that NeuRL leads to better generalization and improved decoding performance compared to supervised approaches. We study the behavior of rats in a response-preparation task and evaluate the performance of NeuRL within simulated inhibition and per-trial behavior prediction. By assigning clear functional roles to defined neuronal populations our approach offers a new interpretation tool for complex neuronal data with testable predictions. In per-trial behavior prediction, our approach furthermore improves accuracy by up to 15% compared to traditional methods.