Safe Reinforcement Learning Using Black-Box Reachability AnalysisShow others and affiliations
2022 (English)In: IEEE Robotics and Automation Letters, E-ISSN 2377-3766, Vol. 7, no 4, p. 10665-10672Article in journal (Refereed) Published
Abstract [en]
Reinforcement learning (RL) is capable of sophisticated motion planning and control for robots in uncertain environments. However, state-of-the-art deep RL approaches typically lack safety guarantees, especially when the robot and environment models are unknown. To justify widespread deployment, robots must respect safety constraints without sacrificing performance. Thus, we propose a Black-box Reachability-based Safety Layer (BRSL) with three main components: (1) data-driven reachability analysis for a black-box robot model, (2) a trajectory rollout planner that predicts future actions and observations using an ensemble of neural networks trained online, and (3) a differentiable polytope collision check between the reachable set and obstacles that enables correcting unsafe actions. In simulation, BRSL outperforms other state-of-the-art safe RL methods on a Turtlebot 3, a quadrotor, a trajectory-tracking point mass, and a hexarotor in wind with an unsafe set adjacent to the area of highest reward.
Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2022. Vol. 7, no 4, p. 10665-10672
Keywords [en]
Reinforcement learning, robot safety, task and motion planning
National Category
Computer Sciences Control Engineering Software Engineering
Identifiers
URN: urn:nbn:se:kth:diva-316727DOI: 10.1109/LRA.2022.3192205ISI: 000838665800016Scopus ID: 2-s2.0-85135247894OAI: oai:DiVA.org:kth-316727DiVA, id: diva2:1691494
Note
QC 20220830
2022-08-302022-08-302024-01-17Bibliographically approved