论文标题
自我解释AI作为可解释AI的替代方案
Self-explaining AI as an alternative to interpretable AI
论文作者
论文摘要
解释AI系统做出的决定的能力受到了极大的追捧,尤其是在人类生命处于危险的领域,例如医学或自动驾驶汽车。虽然通常可以通过一些人为理解的规则近似深度神经网络的投入输出关系,但发现双重下降现象的发现表明,这种近似值并不能准确捕获深神经网络起作用的机制。双重下降表明,深神经网络通常是通过在数据点之间平稳插值而不是提取一些高级规则来运行的。结果,在复杂的现实世界数据上训练的神经网络本质上很难解释,如果被要求推断出来,则容易出现失败。为了展示尽管存在这些问题,我们还可以信任AI,我们介绍了自我解释AI的概念。自我解释的AIS能够为每个决定提供人为理解的解释,并为决策和解释提供信心水平。对于这种工作方法,重要的是,解释实际上与决定相关,理想情况下捕获了用于解释的机制。最后,我们认为,重要的是,基于深度学习的系统基于从适用性域分析的技术包含“警告光”,以警告用户是否要求模型推断其训练分布之外。有关此演讲的视频演示,请参见https://www.youtube.com/watch?v=py7pvdcu7wy&watch?
The ability to explain decisions made by AI systems is highly sought after, especially in domains where human lives are at stake such as medicine or autonomous vehicles. While it is often possible to approximate the input-output relations of deep neural networks with a few human-understandable rules, the discovery of the double descent phenomena suggests that such approximations do not accurately capture the mechanism by which deep neural networks work. Double descent indicates that deep neural networks typically operate by smoothly interpolating between data points rather than by extracting a few high level rules. As a result, neural networks trained on complex real world data are inherently hard to interpret and prone to failure if asked to extrapolate. To show how we might be able to trust AI despite these problems we introduce the concept of self-explaining AI. Self-explaining AIs are capable of providing a human-understandable explanation of each decision along with confidence levels for both the decision and explanation. For this approach to work, it is important that the explanation actually be related to the decision, ideally capturing the mechanism used to arrive at the explanation. Finally, we argue it is important that deep learning based systems include a "warning light" based on techniques from applicability domain analysis to warn the user if a model is asked to extrapolate outside its training distribution. For a video presentation of this talk see https://www.youtube.com/watch?v=Py7PVdcu7WY& .