Skip to content

Latest commit

 

History

History
99 lines (77 loc) · 4.88 KB

index.md

File metadata and controls

99 lines (77 loc) · 4.88 KB

Nonlinear Assimilation with Score-based Sequential Langevin Sampling

Zhao Ding1, Chenguang Duan1, Yuling Jiao1, Jerry Zhijian Yang1, Cheng Yuan2 and Pingwen Zhang1,3

Arranged in alphabetical order.

1 Wuhan University, 2 Central China Normal University, 3 Peking University

GitHub arXiv

Brief

We propose nonlinear assimilation method called score-based sequential Langevin sampling (SSLS) within a Bayesian recursive framework.

Problem Setup

Consider the following system: $$ \begin{align*} \mathbf{X}^{k} & =f_{k-1}(\mathbf{X}^{k-1}, \eta^{k-1}), \quad k > 1, \ \mathbf{Y}^{k} & =g_{k}(\mathbf{X}^{k}, \xi^{k}), \quad k \geq 1, \end{align*} $$ where $\mathbf{X}_k$ is the latent states of interests evolved by $f_k$, and $\mathbf{Y}_k$ is the observations under measurement $g_k$. Here we assume that $\eta^k$ and $\xi^k$ are noises with known distributions.

The goal of Data Assimilation: Combine historical observations with dynamics simulation to provide the best estimate of the current states.

Recursive Bayesian Framework

Our work is carried out under the recursive Bayesian framework described below: $$ \begin{align*} & {\color{blue} {p(\mathbf{x}^k | \mathbf{y}^{[k]})}} \ \propto~ & p(\mathbf{y}^k | \mathbf{x}^k, \mathbf{y}^{[k-1]}) p(\mathbf{x}^k, \mathbf{y}^{[k-1]}) \ \propto~ & \underbrace{p(\mathbf{y}^k | \mathbf{x}^k)}{\text{likelihood}} \underbrace{\int \overbrace{p(\mathbf{x}^k | \mathbf{x}^{k-1})}^{\text{transition}} {\color{blue} \overbrace{ {p(\mathbf{x}^{k-1} | \mathbf{y}^{[k-1]})}}^{\text{last posterior}}} , \mathrm{d} \mathbf{x}^{k-1}}{\text{prior}} \ \propto~ & \underbrace{p(\mathbf{y}^k | \mathbf{x}^k)}{\text{likelihood}} \underbrace{p(\mathbf{x}^{k} | \mathbf{y}^{[k-1]})}{\text{prior}} \end{align*} $$ We maintain an ensemble of particles to estimate the prior and posterior distribution throughout the assimilation process. At each step, the prior samples are obtained by running the dynamics simulation starting from the posterior particles of the last time point.

Langevin Monte Carlo

The posterior score now can be decomposed as the sum of likelihood score and prior score: $$ \underbrace{\nabla \log p (\mathbf{x}^k|\mathbf{y}^{[k]})}\text{score of posterior} = \nabla \log \underbrace{p(\mathbf{y}^k|\mathbf{x}^k)}\text{likelihood} + \underbrace{\nabla \log p(\mathbf{x}^k|\mathbf{y}^{[k-1]})}_\text{score of prior}. $$ The likelihood score can be computed with known measurement model and noises. As for the prior score, we exploit the score matching technique at each time step based on the prior ensemble.

After assembling the posterior score, we can use any Langevin-type sampling method to derive samples from the posterior distribution, starting from the transitioned ensemble from last time step: $$ \mathrm{d} \mathbf{X}_t^k = \nabla \log p(\mathbf{X}_t^k | \mathbf{y}^{[k]}) , \mathrm{d}t + \sqrt{2} , \mathrm{d} \mathbf{B}_t, \ \mathbf{X}_0^k \sim p(\mathbf{x}^k|\mathbf{y}^{[k-1]}), \ t \in [0, \infty). $$

Flow Chart

We provide a flow chart below.

Pseudocode

We provide the python-like pseudocode below.

# start from an initial prior
prior = sample_from_prior()
for i in range(k+1):
    # sliced / implicit / denoising
    prior_score = score_matching(prior)
    # assemble posterior
    posterior_score = lambda x: grad_log_likelihood(x, y[i]) + prior_score(x)
    # any Langevin-type sampling method
    posterior = langevin(prior, posterior_score)
    # dynamics transition to get best guess for next step
    prior = dynamics_transition(posterior)

Numerical Results

Numerical examples demonstrate its outstanding performance in high-dimensional and nonlinear scenarios, as well as in situations with sparse or partial measurements. Please refer to our paper for more results.

Kolmogorov Flow AvgPool Kolmogorov Flow GridMask Kolmogorov Flow CenterMask

How to Cite

If you find our work useful for your research, please consider citing

@misc{ding2024nonlinearassimilationscorebasedsequential,
      title={Nonlinear Assimilation with Score-based Sequential Langevin Sampling}, 
      author={Zhao Ding and Chenguang Duan and Yuling Jiao and Jerry Zhijian Yang and Cheng Yuan and Pingwen Zhang},
      year={2024},
      eprint={2411.13443},
      archivePrefix={arXiv},
      primaryClass={math.NA},
      url={https://arxiv.org/abs/2411.13443}, 
}