Contractivity of neural ODEs: an eigenvalue optimization problem
Nicola Guglielmi,
Arturo De Marinis,
Anton Savostianov,
Francesco Tudisco,
Mathematics of Computation (In press),
(2024)
Abstract
We propose a novel methodology to solve a key eigenvalue optimization problem which arises in the contractivity analysis of neural ODEs. When looking at contractivity properties of a one layer weight-tied neural ODE ˙u(t)=σ(Au(t)+b) (with u,b∈Rn, A is a given n×n matrix, σ:R→R+ denotes an activation function and for a vector z∈Rn, σ(z)∈Rn has to be interpreted entry-wise), we are led to study the logarithmic norm of a set of products of type DA, where D is a diagonal matrix such that diag(D)∈σ′(Rn). Specifically, given a real number c (usually c=0), the problem consists in finding the largest positive interval χ⊆[0,∞) such that the logarithmic norm μ(DA)≤c for all diagonal matrices D with Dii∈χ. We propose a two-level nested methodology: an inner level where, for a given χ, we compute an optimizer D⋆(χ) by a gradient system approach, and an outer level where we tune χ so that the value c is reached by μ(D⋆(χ)A). We extend the proposed two-level approach to the general multilayer, and possibly time-dependent, case ˙u(t)=σ(Ak(t)…σ(A1(t)u(t)+b1(t))…+bk(t)) and we propose several numerical examples to illustrate its behaviour, including its stabilizing performance on a one-layer neural ODE applied to the classification of the MNIST handwritten digits dataset.
Please cite this paper as:
@article{guglielmi2024contractivity,
title={Contractivity of neural ODEs: an eigenvalue optimization problem},
author={Guglielmi, Nicola and De Marinis, Arturo and Savostianov, Anton and Tudisco, Francesco},
journal={arXiv:2402.13092},
year={2024}
}
Links:
arxiv
Keywords:
neural ode
deep learning
neural networks
adversarial attacks
nonlinear eigenvalues
eigenvalue optimization