论文标题
贝叶斯网:将不确定性纳入基于图像的分类任务的神经网络中
BayesNetCNN: incorporating uncertainty in neural networks for image-based classification tasks
论文作者
论文摘要
自动算法提出的信任预测的意愿是许多域中的关键。但是,许多深层体系结构只能在没有相关不确定性的情况下制定预测。在本文中,我们提出了一种将标准神经网络转换为贝叶斯神经网络的方法,并通过对每个正向通行证类似于原始网络进行采样不同的网络来估计预测的可变性。我们将方法与基于可调拒绝的方法相结合,该方法仅采用数据集的比例,该数据集的分数能够以低于用户集阈值的不确定性进行分类。我们在阿尔茨海默氏病患者的大量大脑图像中测试了我们的模型,在那里我们仅根据形态计量学图像来解决与健康对照患者的歧视。我们证明了将估计的不确定性与基于拒绝的方法结合在一起如何将分类精度从0.86提高到0.95,同时保留了75%的测试集。此外,该模型可以根据过度不确定性选择建议进行手动评估的情况。我们认为,能够估计预测的不确定性,以及可以调节网络行为的工具,以使用户被告知(和舒适)可以代表用户合规性方向的关键步骤,并更容易将深度学习工具集成到人类操作员当前执行的日常任务中。
The willingness to trust predictions formulated by automatic algorithms is key in a vast number of domains. However, a vast number of deep architectures are only able to formulate predictions without an associated uncertainty. In this paper, we propose a method to convert a standard neural network into a Bayesian neural network and estimate the variability of predictions by sampling different networks similar to the original one at each forward pass. We couple our methods with a tunable rejection-based approach that employs only the fraction of the dataset that the model is able to classify with an uncertainty below a user-set threshold. We test our model in a large cohort of brain images from Alzheimer's Disease patients, where we tackle discrimination of patients from healthy controls based on morphometric images only. We demonstrate how combining the estimated uncertainty with a rejection-based approach increases classification accuracy from 0.86 to 0.95 while retaining 75% of the test set. In addition, the model can select cases to be recommended for manual evaluation based on excessive uncertainty. We believe that being able to estimate the uncertainty of a prediction, along with tools that can modulate the behavior of the network to a degree of confidence that the user is informed about (and comfortable with) can represent a crucial step in the direction of user compliance and easier integration of deep learning tools into everyday tasks currently performed by human operators.