Sense-Assess-eXplain (SAX): Building trust in autonomous vehicles in challenging real-world driving scenarios

Understanding the decisions taken by an autonomous machine is key to building public trust in robotics and autonomous systems (RAS). This project will design, develop, and demonstrate fundamental AI technologies in real-world applications to address this issue of explainability.

The aim of the project is to build robots, or autonomous vehicles, that can:

  • sense and fully understand their environment,
  • assess their own capabilities,
  • provide causal explanations for their own decisions

Motivated by the question: “How can we develop autonomous vehicles that can explain the decisions they take?”

In on-road and off-road driving scenarios, ORI are studying the requirements of explanations for key stakeholders (users, system developers, regulators). These requirements will inform the development of the algorithms that will generate the causal explanations.

The work will focus on scenarios in which the performance of traditional sensors (e.g cameras) significantly degrades or completely fails (e.g. in harsh weather conditions). The project will develop methods that can assess the performance of perception systems and adapt to environmental changes by switching to another sensor model or a different sensor modality. For the latter, alternative sensing devices will be investigated (incl. radar and acoustic sensors) which can guarantee robust perception in situations when traditional sensors fail.

ORI Investigators: Paul Newman and Lars Kunze
Project Website: https://www.york.ac.uk/assuring-autonomy/projects/sax/