论文标题
通过动作空间对抗训练来鲁棒性增强学习剂
Robustifying Reinforcement Learning Agents via Action Space Adversarial Training
论文作者
论文摘要
在现代社会的各个部门(例如运输,工业和电网)中,采用机器学习(ML)的网络物理系统(CPS)越来越普遍。深度加强学习(DRL)的最新研究表明,其在各种数据驱动的决策和控制应用中的好处。随着对支持ML的系统的依赖,必须研究这些系统在恶意状态和执行器攻击下的性能。传统控制系统采用弹性/耐故障控制器,通过通过错误观察纠正系统来对抗这些攻击。但是,在某些应用中,弹性控制器可能不足以避免灾难性失败。理想情况下,在这些方案中,强大的方法更有用,因为系统本质上是强大的(通过设计)对对抗性攻击。尽管强大的控制有悠久的发展历史,但强大的ML是一个新兴的研究领域,已经证明了其相关性和紧迫性。但是,尽管用于控制应用程序的ML(特别是RL)模型同样容易受到对抗性攻击的影响,但大多数强大的ML研究都集中于感知任务,而不是决策和控制任务。在本文中,我们表明,最初容易受到动作空间扰动(例如执行器攻击)的表现良好的DRL代理,可以通过对抗训练来鲁棒。
Adoption of machine learning (ML)-enabled cyber-physical systems (CPS) are becoming prevalent in various sectors of modern society such as transportation, industrial, and power grids. Recent studies in deep reinforcement learning (DRL) have demonstrated its benefits in a large variety of data-driven decisions and control applications. As reliance on ML-enabled systems grows, it is imperative to study the performance of these systems under malicious state and actuator attacks. Traditional control systems employ resilient/fault-tolerant controllers that counter these attacks by correcting the system via error observations. However, in some applications, a resilient controller may not be sufficient to avoid a catastrophic failure. Ideally, a robust approach is more useful in these scenarios where a system is inherently robust (by design) to adversarial attacks. While robust control has a long history of development, robust ML is an emerging research area that has already demonstrated its relevance and urgency. However, the majority of robust ML research has focused on perception tasks and not on decision and control tasks, although the ML (specifically RL) models used for control applications are equally vulnerable to adversarial attacks. In this paper, we show that a well-performing DRL agent that is initially susceptible to action space perturbations (e.g. actuator attacks) can be robustified against similar perturbations through adversarial training.