Skip to content

Latest commit

 

History

History
82 lines (47 loc) · 3.38 KB

theory.rst

File metadata and controls

82 lines (47 loc) · 3.38 KB

Theory

Inverse Problem

Considering a time-independent system, it can be prescribed as the following model:


x = [θ, α]

where x is state vector, θ is the state variables characterizing the system behaviors and α represents the state augmentation parameters for system design.

Then the stochastic formulation of state vector and observation can be constructed for the system by adding an addictive random term in deterministic formulation as following:


x = x0 + η


y = yo + ϵ

where η is the random perturbation on the initial state vector x0, yo is the available observation of the system. ϵ is random observation error. With random state vector and observation, we can leverage the Bayes's Theorem to get a posterior distribution by combining the prior distribution and data distribution which is called inverse problem.


p(x ∣ y) ∝ p(x)p(y ∣ x)

Thus the problem can be solved with data assimilation by finding state vector to maximize a posterior(MAP).

Data Assimilation

Data assimilation can be sorted by variational method le1986variational and Kalman filter kalman1960new. The variational method use the optimal control theory berkovitz2013optimal optimal to reduce the misfit between observation and model realization in observed space, while Kalman filter is originated directly from the Bayesian formulation. However, the variational method need much efforts on the adjoint model and the Kalman filter is quite costly to estimate the statics of the state vector especially for high-dimension problem. To address these issues, the trending approach is to introduce ensemble technique, and many ensemble-based methods have been proposed.

This toolbox is focused on ensemble-based data assimilation methods.

Ensemble Kalman Filtering (EnKF)

In EnKF, the prior statistics is estimated by ensemble Monte Carlo sampling and each sample in kth DA step can be defined as:


xk(j) = [θk(j), αk(j)]  1 ≤ j ≤ Nen

where Nen is the ensemble size.

In bayes, MAP is equivalent to the determination of a variance minimizing analysis. Accordingly, the matrix is updated as:


xk + 1(j) = xk(j) + K(y(j) − H(xk(j)))

with gain matrix K = PHT(HPHT + C) − 1, $P = \frac{1}{N_{en}-1}X'X'^T$, $X'=(x^{(1)}-x^{(e)},x^{(2)}-x^{(e)},\dot,x^{(N_{en})}-x^{(e)})$ where H is operator that map from the state space to observation space, C is observation covariance error.

The procedure of EnKF can be summarized as follows:

  1. Give a first guessed or prior state vector x(e), and prescribe the prior and observation statistics respectively;
  2. Realize Nen initial samples {x(j)}j = 1N around x(e);
  3. Map the state vector to observation space by solving the RANS equation based on each sample and obtain HX;
  4. Obtain gain matrix K and update the prior distribution based on stateupdate_EnKF;
  5. Return to step 3 until further minimization cannot be achieved.

For more information, please refer to xiao2016

References

theory.bib