Safety Certificate against Latent Variables with Partially Unidentifiable Dynamics

Published: 01 May 2025, Last Modified: 16 Aug 2025ICML 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: We propose a causal-RL–driven probabilistic safety certificate that accounts for latent-variable–induced distribution shifts by enforcing an invariance condition in probability space, enabling persistent feasible actions and long-term risk control.
Abstract: Existing control techniques often assume access to complete dynamics or perfect simulators with fully observable states, which are necessary to verify whether the system remains within a safe set (forward invariance) or safe actions are persistently feasible at all times. However, many systems contain latent variables that make their dynamics partially unidentifiable or cause distribution shifts in the observed statistics between offline and online data, even when the underlying mechanistic dynamics are unchanged. Such “spurious” distribution shifts can break many techniques that use data to learn system models or safety certificates. To address this limitation, we propose a technique for designing probabilistic safety certificates for systems with latent variables. A key technical enabler is the formulation of invariance conditions in probability space, which can be constructed using observed statistics in the presence of distribution shifts due to latent variables. We use this invariance condition to construct a safety certificate that can be implemented efficiently in real-time control. The proposed safety certificate can persistently find feasible actions that control long-term risk to stay within tolerance. Stochastic safe control and (causal) reinforcement learning have been studied in isolation until now. To the best of our knowledge, the proposed work is the first to use causal reinforcement learning to quantify long-term risk for the design of safety certificates. This integration enables safety certificates to efficiently ensure long-term safety in the presence of latent variables. The effectiveness of the proposed safety certificate is demonstrated in numerical simulations.
Lay Summary: Many real-world systems, like robots or self-driving cars, operate in environments where not everything can be seen or measured. This hidden information can make it hard to predict how the system will behave or to ensure it stays safe. Most existing safety methods assume we have complete knowledge or perfect simulations of these systems—which isn’t realistic. This paper introduces a new method to help systems stay safe even when some information is missing. This method can quickly compute safe actions while the system is running, helping it avoid risky situations in the long run. This is the first approach that combines safety techniques with a class of machine learning techniques—called causal reinforcement learning—to better understand and manage long-term risks when certain information in the environment is hidden.
Primary Area: Theory->Everything Else
Keywords: Control System, Stochastic System, Safety
Submission Number: 12690
Loading