The Foundation Inference Model Program#

Foundation Inference Models (FIMs) are pretrained deep neural network models that perform zero-shot (or in-context) inference of dynamical systems from noisy time series data, and can easily be fine-tuned to specific tasks.

Our released FIMs include:

Note

Take a look at our tutorials!

Currently, we are working on releasing

  • FIMs for Point Processes (preprint)

  • FIMs for Ordinary Differential Equations (preprint)

  • FIMs for Accelerating Coarse-graining and Equation Discovery (preprint)

Defining FIMs#

A satisfactory FIM should be able to cope with dynamic phenomena of very diverse nature and thus

  • capture various secular and seasonal patterns, as well as non-periodic fluctuations;

  • handle observations from empirical processes of different dimensionalities;

  • process any number of such observations, which might be recorded regularly in time or not; and

  • deal with noise signals of different kinds.

To build such FIMs, we follow a general three-step pretraining strategy (FIM workflows, upper right):

_images/f1.png

Fig. 1 FIM workflows#

  1. Construct a broad prior probability distribution over the space of dynamical systems. This distribution represents our beliefs about the general class of systems we expect to encounter in nature.

  2. Sample dynamical systems from this prior distribution, simulate them, and corrupt the simulated paths to generate a dataset of noisy observations and target dynamical systems. This step defines a supervised, meta-learning task that amortizes the inference process.

  3. Train a FIM to match these observation-(target-)system pairs in a supervised way.

Once (pre)trained

  • FIMs can accurately infer the governing dynamical system in zero-shot (in-context) mode from unseen, noisy, and heterogeneous datasets (FIM workflows, lower left); and

  • FIMs can also rapidly be finetuned to target datasets (FIM workflows, lower right).

FIMs therefore amortize the inference of dynamical systems.

About us: Deep Learning for Scientific Discovery Group#

The Deep Learning for Scientific Discovery (DL4SD) group at the Lamarr Institute is an interdisciplinary team working at the intersection of machine learning, natural language processing, statistical physics, and complexity science.

At DL4SD, our aim is to tackle system identification problems across the natural and social sciences. This means that we develop artificial intelligence methods to infer dynamical systems — such as ordinary, partial, and stochastic differential equations — from noisy, sparse and high-dimensional data. These models deliver interpretable and predictive insights into the underlying mechanisms governing dynamic phenomena.

Ultimately, our goal is to automate the formulation of scientific hypotheses, expressed as mathematical models derived directly from data, to better understand complex dynamic phenomena.

Note

Take a deeper look at our team.

The Problem of System Identification or Dynamical System Inference#

Inferring dynamical systems from data involves two interconnected challenges:

  • representation learning — which involves determining appropriate coarse-grained data representations, and

  • dynamical model fitting — which consists in learning an interpretable set of equations (e.g., ODEs, SDEs) that represent the changes in those representations.

Traditional methods such as symbolic and Gaussian process regression or Neural-ODE/SDE/MJP rely on variational inference to jointly solve both tasks, but its slow convergence and intricate optimization render it impractical for scalable, automated scientific discovery.

The FIM program decouples these learning processes by amortizing the model-fitting step. Indeed, FIMs introduce a form of in-context, simulation-based inference by encoding canonical mathematical models (ODEs, SDEs, MJPs, etc) widely used in science.