Table of Contents
Fetching ...

Learning Certified Neural Network Controllers Using Contraction and Interval Analysis

Akash Harapanahalli, Samuel Coogan, Alexander Davydov

Abstract

We present a novel framework that jointly trains a neural network controller and a neural Riemannian metric with rigorous closed-loop contraction guarantees using formal bound propagation. Directly bounding the symmetric Riemannian contraction linear matrix inequality causes unnecessary overconservativeness due to poor dependency management. Instead, we analyze an asymmetric matrix function $G$, where $2^n$ GPU-parallelized corner checks of its interval hull verify that an entire interval subset $X$ is a contraction region in a single shot. This eliminates the sample complexity problems encountered with previous Lipschitz-based guarantees. Additionally, for control-affine systems under a Killing field assumption, our method produces an explicit tracking controller capable of exponentially stabilizing any dynamically feasible trajectory using just two forward inferences of the learned policy. Using JAX and $\texttt{immrax}$ for linear bound propagation, we apply this approach to a full 10-state quadrotor model. In under 10 minutes of post-JIT training, we simultaneously learn a control policy $π$, a neural contraction metric $Θ$, and a verified 10-dimensional contraction region $X$.

Learning Certified Neural Network Controllers Using Contraction and Interval Analysis

Abstract

We present a novel framework that jointly trains a neural network controller and a neural Riemannian metric with rigorous closed-loop contraction guarantees using formal bound propagation. Directly bounding the symmetric Riemannian contraction linear matrix inequality causes unnecessary overconservativeness due to poor dependency management. Instead, we analyze an asymmetric matrix function , where GPU-parallelized corner checks of its interval hull verify that an entire interval subset is a contraction region in a single shot. This eliminates the sample complexity problems encountered with previous Lipschitz-based guarantees. Additionally, for control-affine systems under a Killing field assumption, our method produces an explicit tracking controller capable of exponentially stabilizing any dynamically feasible trajectory using just two forward inferences of the learned policy. Using JAX and for linear bound propagation, we apply this approach to a full 10-state quadrotor model. In under 10 minutes of post-JIT training, we simultaneously learn a control policy , a neural contraction metric , and a verified 10-dimensional contraction region .

Paper Structure

This paper contains 10 sections, 5 theorems, 22 equations, 1 figure, 1 table, 1 algorithm.

Key Result

Lemma 1

Consider the vector field $f_\pi$ from eq:fpi, let $M(x) = \Theta(x)^T\Theta(x)$ be a Riemannian metric, and $c>0$ be fixed. The following are equivalent: Thus, $X$ is a contraction region if and only if $\mu_2(G(t,x)) \leq 0$ uniformly over $t\in[0,\infty)$, $x\in X$.

Figures (1)

  • Figure 2: Plots (a)-(d) demonstrate how 10 randomly sampled initial conditions (pictured with the $\times$) sampled in the verified contraction region $X$ stabilize to the nominal trajectories pictured in red under the tracking policy from Corollary \ref{['cor:tracking']}. Plot (e) shows the state vs time for 100 randomly sampled initial conditions starting in $X$ for the trefoil knot (d). The nominal trajectory is in red, with the verified contraction region $X$ in blue. While some trajectories leave the verified region, the controller empirically behaves well, bringing them back into the region and stabilizing to the commanded trajectory.

Theorems & Definitions (17)

  • Definition 1: Contraction Region lohmiller_contraction_1998
  • Definition 2: Asymmetric Contraction Matrix
  • Lemma 1: Equivalence
  • proof
  • Example 1: Dependency management
  • Lemma 2: rohn_positive_1994
  • proof
  • Remark 1
  • Theorem 1
  • proof
  • ...and 7 more