Skip to content

maxbeauchamp/4DVARNN-DinAE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

End-to-end learning of variational/energy models and solvers for inverse problems (especially with noisy and irregularly-sampled observations)

Introduction

Associated preprints:

License: CECILL-C license

Copyright IMT Atlantique/OceaniX, contributor(s) : R. Fablet, 21/03/2020

Contact person: ronan.fablet@imt-atlantique.fr This software is a computer program whose purpose is to apply deep learning schemes to dynamical systems and ocean remote sensing data. This software is governed by the CeCILL-C license under French law and abiding by the rules of distribution of free software. You can use, modify and/ or redistribute the software under the terms of the CeCILL-C license as circulated by CEA, CNRS and INRIA at the following URL "http://www.cecill.info". As a counterpart to the access to the source code and rights to copy, modify and redistribute granted by the license, users are provided only with a limited warranty and the software's author, the holder of the economic rights, and the successive licensors have only limited liability. In this respect, the user's attention is drawn to the risks associated with loading, using, modifying and/or developing or reproducing the software by the user in light of its specific status of free software, that may mean that it is complicated to manipulate, and that also therefore means that it is reserved for developers and experienced professionals having in-depth computer knowledge. Users are therefore encouraged to load and test the software's suitability as regards their requirements in conditions enabling the security of their systems and/or data to be ensured and, more generally, to use and operate it in the same conditions as regards security. The fact that you are presently reading this means that you have had knowledge of the CeCILL-C license and that you accept its terms.

Architecture of the code

It will involve a joint learning scheme of operator and solver through a bi-level optimization scheme:

where is the 4DVAR cost function:

and is the loss function to use when using automatic differential tools in the neural network. This loss is defined according to the GradType flag in utils/utils_solver/Compute_Grad.py:

  • GradType == 0: subgradient for prior
  • GradType == 1: true gradient using autograd for prior
  • GradType == 2: true gradient using autograd for prior
  • GradType == 3: true gradient using autograd for prior
  • GradType == 4: true gradient using autograd for prior

According to the OptimType flag, the appropriate gradient-based solver is defined in the corresponding model_GradUpdateX.py for the minimization, i.e.:

  • if OptimType == 0, load utils/utils_solver/model_GradUpdate0.py : Gradient-based minimization using a fixed-step descent
  • if OptimType == 1, load utils/utils_solver/model_GradUpdate1.py : Gradient-based minimization using a CNN using a (sub)gradient as inputs
  • if OptimType == 2, load utils/utils_solver/model_GradUpdate2.py : Gradient-based minimization using a 2D convolutional LSTM using a (sub)gradient as inputs defined in ConvLSTM2d.py

The NN-based dynamical operator is defined through flagAEtype:

  • utils/utils_nn/ConvAE.py: Convolutional auto-encoder
  • utils/utils_nn/GENN.py: Gibbs-Energy-based NN for using a 2d convolution ResNet architecture (utils/utils_nn/ResNetConv2d.py) and a 2d convolution Constrained layer with zero-weight at central point (utils/utils_nn/ConstrainedConv2d.py)

The general model for optimizing both parameters in and is defined in utils/utils_solver/Model_4DVarNN_GradFP.py. It uses a set of (NProjection) fixed-point iterations to fill in the missing-data in the output of the NN; then, a predefined number (NGradIter) Gradient-based iterations are involved to converge through the best operator and solver for the bi-level optimization scheme.

Results

Below is an illustration of the results obtained on the daily velocity SSH field when interpolating pseudo irregular and noisy observations (top-right panels) corresponding to along-track nadir (left) with additional pseudo wide-swath SWOT observations (right) built from an idealized groundtruth (top-left panels) with state-of-the-art optimal interpolation (bottom-left panels) and the newly proposed end-to-end learning approach (bottom-right panels):

Nadir only Nadir+SWOT
Farmers Market Finder Demo Farmers Market Finder Demo

About

End-to-end learning for Variational Assimilation models and solvers

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published