Home | People | Research | Publications | Software | Links | Youtube Channel | News | Contact   

LEAD: Learning-Enabled Assistive Driving: Formal Assurances during Operation and Training

Current driver-assist systems are the epitomy of CPS as they integrate the physical and digital domains by incorporating a myriad of Electronic Control Units (ECUs) running sophisticated planning and control algorithms, so much so that they are aptly considered as “computers on wheels”. With the proliferation of data-driven algorithms,
future self-driving vehicles will incorporate even more advanced technologies, such as four-wheel independent steering, drive-by-wire, brake-by-wire, perception modules, autonomous navigation, lane-change, platooning, etc. and will be able to solve many critical tasks such as perception, decision making, and adaptive control in unstructured environments and move towards complete autonomy. However, due to the fragility of present machine learning algorithms in adapting to unseen environments or unaccounted contextual situations, human intervention is still necessary. What is not clear is how to align the decisions taken by the vehicle’s autonomous or active safety control system with the human’s intentions, goals, and skill set. This is paramount for passenger vehicles operating in an “information-rich” world that require the smooth and safe interaction between machine, computer, and human at all levels: the driver, the vehicle, and the traffic. While the modeling and control aspects at each level have received their due attention in the literature, we argue that the interaction between these three levels demarcates the cyber-physical nature of the problem

Current driving autonomy solutions leverage advanced machine learning algorithms to perform challenging pattern recognition, planning and high-level decision tasks. In particular, reinforcement learning (RL) has proved successful in learning strategies and tasks, often out-performing humans. At the same time, these RL algorithms must align the agent’s goals with the human driver without sacrificing safety. Such interactions are vital for safe operation, yet they have not been deeply explored thus far.

 

There is currently a need to quantify the impact of the human driver within the autonomy loop, both from an individual experiential perspective, as well as in terms of safety. In this research, we propose to increase the performance and safety guarantees of deep neural network architectures operating within a feedback loop that includes the driver by: a) using redundant architectures that blend model-free and model-based processing pipelines, and b) by adding safety guarantees both during training and during execution by leveraging recent advances of formal methods for safety-critical applications..

The proposed effort makes fundamental contributions to the challenging problem of safe operation of (deep) neural network-based learning architectures both during training and during execution. Our methodology is based on four key novel ingredients that guide our technical approach. The first ingredient is the introduction of a novel deep NN architecture that combines both model-based representation sensor measurements to enhance robustness during training. The second ingredient is the incorporation of signal temporal logic (STL) specifications within the proposed DNN architecture via the introduction of a bi-directional recurrent neural network (RNN) to ensure safety during training of learning-enabled advanced driver assist systems (ADAS), while also considering the driver’s habits and driving skills. The third ingredient is the development of new learning-from-demonstration techniques that enable population-based training of controllers to perform a context- or task-specific blending of individual, heterogeneous driver policies. Finally, we utilize efficient reachable set approximations to formulate an optimization-based runtime assurance (RTA) mechanism to ensure the satisfaction of the STL specifications during execution via the use of the theory of mixed monotonicity for efficient reachable set calculations of NNs.

The project is in collaboration with Sam Coogan and Matthew Gombolay.

Please also visit our VIP project website.

Sponsors

This project is supported by NSF.

Selected Publications

  • TBA
  • TBA
 

Home | People | Research | Publications | Software | Links | Youtube Channel | News | Contact