Goal-oriented physics-constrained machine learning for chaotic systems

When:
21/05/2019 – 22/05/2019 all-day
2019-05-21T02:00:00+02:00
2019-05-22T02:00:00+02:00

Annonce en lien avec l’Action/le Réseau : aucun

Laboratoire/Entreprise : LIMSI – CNRS
Durée : 36 mois
Contact : mathelin@limsi.fr
Date limite de publication : 2019-05-21

Contexte :
The rise of machine learning techniques has led to major efforts to enhance, and sometimes substitute, physics solvers. Of particular interest for scientific purposes is the modeling of the behavior of a system from limited observations. In particular, one often needs to learn faithful and reliable governing equations accurately describing the time-evolution of the system under consideration.

The class of models generically described as “neural networks” has seen a spectacular renewed interest in the last decade and has now become the gold standard in many applications. In particular, deep networks have often demonstrated impressive performance. In a recent work, a deep neural network architecture (Reservoir Computing) has been used to learn a model for the chaotic Kuramoto-Sivashinsky system. The resulting model is able to accurately predict the future state of the system up to about 6 Lyapunov exponent times.

This remarkable result rises several largely open questions, also prototypical of a large class of learning situations:
• What are the reasons why this deep neural computing allows such a performance, well beyond the usual Lyapunov time limit? How specific is the performance to the particular architecture of this network?
• In a broad sense, what is the best quantity to learn for achieving good general performance? More precisely, what is the good metric to use as the training criterion to improve generalizability of the model learned? In a dual view, what is a relevant nonlinear transform to pre-process the data so that it achieves a good validation error?
• In many situations, expertise knowledge is available on the system under consideration. How to make use of this expertise and reduce the dimension of the learning space by introducing relevant constraints such as symmetries, invariants, causality, etc.?
• In situations of active learning when the samples can be chosen, how to benefit from past knowledge to improve the sampling scheme? How to adapt the design of experiment and explore the system in an efficient way, given the objective function?

Sujet :
The objectives of this thesis will be to improve the learning process of a data-driven approach by addressing the questions and limitations above. Specifically, different points will be addressed, including but not limited to:

1. Choice of the best cost function to get optimal results. Cost functions are typically defined in terms of integral (L2) norm of the misfit residual. This choice is not always necessary and one may want to employ other metrics. We would like to use a physics-based approach to decide which feature deserves particular attention. For instance, one may need to ensure a good learning in terms of Fourier spectrum, Lyapunov exponents, wavelet structure, optimal transport-related norms (e.g., Wasserstein distance), etc. We will first analyze how a standard L2 learning is relevant for other metrics. In a second step, these alternative metrics will be directly employed as objective functions in a goal-oriented approach.

2. Many key questions will be at the center of the effort:
How to sample the system under consideration in the most efficient way? How to most quickly discover its natural measure and adapt the sampling scheme accordingly? How to explore around the neutral manifold (when it exists) in useful directions? What is the minimal number of samples one needs to learn a model of an attractor of a given dimension?

Profil du candidat :
This project is strongly multidisciplinary (applied maths, statistics, dynamical systems, fluid mechanics, data analysis, high-performance computing applied to engineering), at the edge of fundamental and applied research.
The successful candidate has a strong background in at least one of the following fields: statistics, applied mathematics, machine learning. He/She is also a team-oriented person, with good communication and writing skills.

Formation et compétences requises :
Programming skills (Python/PyTorch, Julia or Matlab) are clearly a plus.

Adresse d’emploi :
LIMSI – CNRS
Campus Universitaire d’Orsay
91405 Orsay cedex
France

Document attaché : Sujet_de_these_19_anglais_short.pdf