Processing math: 100%
Francesco Tudisco

Contractivity of neural ODEs: an eigenvalue optimization problem

Nicola Guglielmi, Arturo De Marinis, Anton Savostianov, Francesco Tudisco,
Mathematics of Computation (In press), (2024)

Abstract

We propose a novel methodology to solve a key eigenvalue optimization problem which arises in the contractivity analysis of neural ODEs. When looking at contractivity properties of a one layer weight-tied neural ODE ˙u(t)=σ(Au(t)+b) (with u,bRn, A is a given n×n matrix, σ:RR+ denotes an activation function and for a vector zRn, σ(z)Rn has to be interpreted entry-wise), we are led to study the logarithmic norm of a set of products of type DA, where D is a diagonal matrix such that diag(D)σ(Rn). Specifically, given a real number c (usually c=0), the problem consists in finding the largest positive interval χ[0,) such that the logarithmic norm μ(DA)c for all diagonal matrices D with Diiχ. We propose a two-level nested methodology: an inner level where, for a given χ, we compute an optimizer D(χ) by a gradient system approach, and an outer level where we tune χ so that the value c is reached by μ(D(χ)A). We extend the proposed two-level approach to the general multilayer, and possibly time-dependent, case ˙u(t)=σ(Ak(t)σ(A1(t)u(t)+b1(t))+bk(t)) and we propose several numerical examples to illustrate its behaviour, including its stabilizing performance on a one-layer neural ODE applied to the classification of the MNIST handwritten digits dataset.

Please cite this paper as:

@article{guglielmi2024contractivity,
  title={Contractivity of neural ODEs: an eigenvalue optimization problem},
  author={Guglielmi, Nicola and  De Marinis, Arturo and Savostianov, Anton and Tudisco, Francesco},
  journal={arXiv:2402.13092},
  year={2024}
}

Links: arxiv

Keywords: neural ode deep learning neural networks adversarial attacks nonlinear eigenvalues eigenvalue optimization