论文标题
预期的功利主义
Expected Utilitarianism
论文作者
论文摘要
我们希望人工智能(AI)有益。这是对AI研究的大多数态度的基础假设。我们希望AI对人类“好”。我们希望它能提供帮助,而不是阻碍人类。然而,从理论上讲,这在实践中究竟需要什么,这并不是立即显而易见的。从理论上讲,这一声明性陈述巧妙地暗示着对结果主义伦理的承诺。实际上,一些更有前途的机器学习技术来创建强大的人工智能,甚至可能是人工通用情报(AGI)也致力于一种功利主义的形式。在这两个维度中,如果伦理假设未明确和清晰,则有益AI运动的逻辑实际上可能不会在狭窄的应用或AGI形式中创建“有益的AI”。 此外,由于强化学习(RL)可能是该领域机器学习的重要技术,因此询问在AI中如何询问RL在特定类型的结果主义推理中的RL走私:尤其是享乐主义行为的野蛮形式。由于数学逻辑将一个最大化函数提交,因此结果是AI将不可避免地寻求越来越多的奖励。我们有两个结论。首先,如果一个人认为有益的AI是道德AI,那么人们将致力于提出“福利”的框架,这是最大的数字最大的好处。其次,如果AI依靠RL,那么它本身,环境和其他代理商的理由将是通过实用性的道德。实际上,这种主张实际上可能对人类有益。
We want artificial intelligence (AI) to be beneficial. This is the grounding assumption of most of the attitudes towards AI research. We want AI to be "good" for humanity. We want it to help, not hinder, humans. Yet what exactly this entails in theory and in practice is not immediately apparent. Theoretically, this declarative statement subtly implies a commitment to a consequentialist ethics. Practically, some of the more promising machine learning techniques to create a robust AI, and perhaps even an artificial general intelligence (AGI) also commit one to a form of utilitarianism. In both dimensions, the logic of the beneficial AI movement may not in fact create "beneficial AI" in either narrow applications or in the form of AGI if the ethical assumptions are not made explicit and clear. Additionally, as it is likely that reinforcement learning (RL) will be an important technique for machine learning in this area, it is also important to interrogate how RL smuggles in a particular type of consequentialist reasoning into the AI: particularly, a brute form of hedonistic act utilitarianism. Since the mathematical logic commits one to a maximization function, the result is that an AI will inevitably be seeking more and more rewards. We have two conclusions that arise from this. First, is that if one believes that a beneficial AI is an ethical AI, then one is committed to a framework that posits 'benefit' is tantamount to the greatest good for the greatest number. Second, if the AI relies on RL, then the way it reasons about itself, the environment, and other agents, will be through an act utilitarian morality. This proposition may, or may not, in fact be actually beneficial for humanity.