arXiv Analytics

Sign in

arXiv:2308.02594 [cs.LG]AbstractReferencesReviewsResources

SMARLA: A Safety Monitoring Approach for Deep Reinforcement Learning Agents

Amirhossein Zolfagharian, Manel Abdellatif, Lionel C. Briand, Ramesh S

Published 2023-08-03Version 1

Deep reinforcement learning algorithms (DRL) are increasingly being used in safety-critical systems. Ensuring the safety of DRL agents is a critical concern in such contexts. However, relying solely on testing is not sufficient to ensure safety as it does not offer guarantees. Building safety monitors is one solution to alleviate this challenge. This paper proposes SMARLA, a machine learning-based safety monitoring approach designed for DRL agents. For practical reasons, SMARLA is designed to be black-box (as it does not require access to the internals of the agent) and leverages state abstraction to reduce the state space and thus facilitate the learning of safety violation prediction models from agent's states. We validated SMARLA on two well-known RL case studies. Empirical analysis reveals that SMARLA achieves accurate violation prediction with a low false positive rate, and can predict safety violations at an early stage, approximately halfway through the agent's execution before violations occur.

Related articles: Most relevant | Search more
arXiv:2104.03154 [cs.LG] (Published 2021-04-07)
Improving Robustness of Deep Reinforcement Learning Agents: Environment Attacks based on Critic Networks
arXiv:2107.00956 [cs.LG] (Published 2021-07-02)
SocialAI: Benchmarking Socio-Cognitive Abilities in Deep Reinforcement Learning Agents
arXiv:1904.06312 [cs.LG] (Published 2019-04-12)
Let's Play Again: Variability of Deep Reinforcement Learning Agents in Atari Environments