论文标题

以负责任的方式建立AI的道德规范

On the Ethics of Building AI in a Responsible Manner

论文作者

Shalev-Shwartz, Shai, Shammah, Shaked, Shashua, Amnon

论文摘要

当人类设计师向AI学习者指定的目标与潜在的灾难性结果之间存在差异时,就会出现AI对准问题,而这并不能反映人类设计师真正想要的东西。我们认为,没有区分战略和不可知的未对准的AI一致性的形式主义是没有用的,因为它认为所有技术是不安全的。我们提出了对战略性对准的定义,并证明当今实践中使用的大多数机器学习算法都不会遭受战略与ai对准问题的困扰。但是,在不仔细的情况下,当今的技术可能会导致战略错位。

The AI-alignment problem arises when there is a discrepancy between the goals that a human designer specifies to an AI learner and a potential catastrophic outcome that does not reflect what the human designer really wants. We argue that a formalism of AI alignment that does not distinguish between strategic and agnostic misalignments is not useful, as it deems all technology as un-safe. We propose a definition of a strategic-AI-alignment and prove that most machine learning algorithms that are being used in practice today do not suffer from the strategic-AI-alignment problem. However, without being careful, today's technology might lead to strategic misalignment.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源