An Abstraction-based Technique for Safe Reinforcement Learning (EPSRC NIA Project) | |
---|---|
OverviewAutonomous agents learning to act in unknown environments have been attracting research interest due to their wider implications for AI, as well as for their applications, including robotics, network optimisation, and resource allocation. Currently, one of the most successful approaches is reinforcement learning (RL). However, to learn how to act, agents are required to explore the environment, which in safety-critical scenarios means that they might take dangerous actions, possibly harming themselves or even putting human lives at risk. Consequently, reinforcement learning is still rarely used in real-world applications, where multiple safety-critical constraints need to be satisfied simultaneously. To alleviate this problem, RL algorithms are being combined with formal verification techniques to ensure safety in learning. Indeed, formal methods are nowadays routinely applied to the specification, design, and verification of complex systems, as they allow to obtain proof-like certification of their correct and safe behaviour, which is meant to be intelligible to system engineers and human users alike. These desirable features have fostered the adoption of formal methods for the verification of general AI systems, which has variously been called safe, verifiable, trustworthy AI. Still, the application of formal methods to AI systems raises significant new challenges, also due to the “black-box” nature of most machine learning algorithms currently used. Specific to the application of formal methods to RL, we identify two main shortcomings with present approaches, which will be tackled in this project:
|