SOTAVerified

Interpretable Safety Validation for Autonomous Vehicles

2020-04-14Code Available0· sign in to hype

Anthony Corso, Mykel J. Kochenderfer

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

An open problem for autonomous driving is how to validate the safety of an autonomous vehicle in simulation. Automated testing procedures can find failures of an autonomous system but these failures may be difficult to interpret due to their high dimensionality and may be so unlikely as to not be important. This work describes an approach for finding interpretable failures of an autonomous system. The failures are described by signal temporal logic expressions that can be understood by a human, and are optimized to produce failures that have high likelihood. Our methodology is demonstrated for the safety validation of an autonomous vehicle in the context of an unprotected left turn and a crosswalk with a pedestrian. Compared to a baseline importance sampling approach, our methodology finds more failures with higher likelihood while retaining interpretability.

Tasks

Reproductions