Research & Papers

Temporal Logic Control of Nonlinear Stochastic Systems with Online Performance Optimization

Novel control method provides multiple safe policies, enabling real-time performance optimization without breaking safety guarantees.

Deep Dive

A research team from TU Delft and the University of Oxford has published a novel control framework for autonomous systems. The paper, 'Temporal Logic Control of Nonlinear Stochastic Systems with Online Performance Optimization,' addresses a key limitation in safety-critical control. Traditional methods create a single, safe control policy by abstracting a complex, nonlinear stochastic system into a simpler Interval Markov Decision Process (IMDP). While this guarantees safety, it leaves no room for real-time optimization of other objectives like energy efficiency or speed.

The new technique, developed by Alessandro Riccardi, Thom Badings, Luca Laurenti, Alessandro Abate, and Bart De Schutter, fundamentally changes this paradigm. Instead of outputting one policy, their IMDP abstraction yields a *set* of policies, all of which are mathematically verified to satisfy the required safety specifications with a guaranteed minimum probability. This creates a 'playground' of safe options. An online controller, such as Model Predictive Control (MPC), can then dynamically select the best policy from this set to minimize a separate cost function, optimizing for performance in real-time.

Experiments demonstrate that this approach achieves significantly better control performance—such as lower energy consumption—compared to state-of-the-art single-policy abstraction techniques. The trade-off is a small, quantifiable degradation in the theoretical safety guarantees, but the system remains provably safe within the bounds of the verified policy set. This breakthrough bridges the gap between rigorous safety verification and practical, adaptive performance, a major hurdle for deploying autonomous systems in unpredictable real-world environments.

Key Points
  • Generates a verified *set* of safe control policies from an IMDP abstraction, not just a single policy.
  • Enables real-time performance optimization (e.g., energy use) via MPC while staying within the safe policy bounds.
  • Experimental results show better control performance than single-policy methods with only a small guarantee degradation.

Why It Matters

Enables autonomous robots and vehicles to dynamically optimize for efficiency and speed without compromising their provable safety guarantees.