论文标题

道德不确定性下的强化学习

Reinforcement Learning Under Moral Uncertainty

论文作者

Ecoffet, Adrien, Lehman, Joel

论文摘要

机器学习的一个雄心勃勃的目标是创造具有道德行为的代理人:遵守人类道德规范的能力将大大扩展自主代理可以实际和安全地部署的自主代理的背景,例如完全自主的车辆将遇到充电的道德决定,使其部署变得复杂。尽管可以通过在特定的道德理论(例如功利主义)下奖励正确行为来训练道德代理人,但人们对道德本质仍然存在普遍分歧。认识到这种分歧,道德哲学的最新工作表明,道德行为需要在道德不确定性下行事,即在表现出一个人的信任性在几种合理的道德理论中分裂时要考虑到。本文将这种见解转化为强化学习领域,提出了两种培训方法,这些培训方法在竞争的逃避者中实现了不同的观点,并在简单环境中训练代理商在道德不确定性下采取行动。结果说明了(1)这种不确定性如何帮助遏制从承诺到单一理论的极端行为,以及(2)试图在RL中尝试基础道德哲学引起的几种技术并发症(例如,如何在两个竞争但无与伦比的奖励功能之间进行原则性的权衡)。目的是催化朝着道德同能力的推动者促进进步,并强调RL为道德哲学的计算基础做出贡献的潜力。

An ambitious goal for machine learning is to create agents that behave ethically: The capacity to abide by human moral norms would greatly expand the context in which autonomous agents could be practically and safely deployed, e.g. fully autonomous vehicles will encounter charged moral decisions that complicate their deployment. While ethical agents could be trained by rewarding correct behavior under a specific moral theory (e.g. utilitarianism), there remains widespread disagreement about the nature of morality. Acknowledging such disagreement, recent work in moral philosophy proposes that ethical behavior requires acting under moral uncertainty, i.e. to take into account when acting that one's credence is split across several plausible ethical theories. This paper translates such insights to the field of reinforcement learning, proposes two training methods that realize different points among competing desiderata, and trains agents in simple environments to act under moral uncertainty. The results illustrate (1) how such uncertainty can help curb extreme behavior from commitment to single theories and (2) several technical complications arising from attempting to ground moral philosophy in RL (e.g. how can a principled trade-off between two competing but incomparable reward functions be reached). The aim is to catalyze progress towards morally-competent agents and highlight the potential of RL to contribute towards the computational grounding of moral philosophy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源