Abstract:
Reinforcement learning (RL) is in essence a trial-and-error process which involves exploratory actions. These explorations can lead to system constraint violations and ph...Show MoreMetadata
Abstract:
Reinforcement learning (RL) is in essence a trial-and-error process which involves exploratory actions. These explorations can lead to system constraint violations and physical system damages, impeding RL's use in many real-world engineered systems. In this paper, we develop a safe RL framework that integrates model-free learning with model-based safety supervision to bridge the gap. We exploit the underlying system dynamics and safety-related constraints to construct a safety set using recursive feasibility techniques. We then integrate the safety set in RL's exploration to guarantee safety while simultaneously preserving exploration efficiency by using the hit-and-run sampling. We design a novel efforts-to-remain-safe penalty to effectively guide RL to learn system constraints. We apply the proposed safe RL framework to the active suspension system in which actuation and state constraints are present due to ride comfort, road handling, and actuation limits. We show that the developed safe RL is able to learn a safe control policy safely while outperforming a nominal controller.
Date of Conference: 19-21 August 2019
Date Added to IEEE Xplore: 05 December 2019
ISBN Information: