yongquan-qu / SLAMS

SLAMS: Score-based Latent Assimilation in Multimodal Setting

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

SLAMS: Score-based Latent Assimilation in Multimodal Setting

🎊🎊 We won the best student paper award at CVPR EarthVision 2024!! 🎊🎊

πŸ“š Paper: https://arxiv.org/abs/2404.06665

πŸ” Overview: We recast data assimilation in a multimodal setting using a deep generative framework. In particular, we implement a latent score-based diffusion model. We project heterogeneous states and observations into a unified latent space where the forward and reverse conditional diffusion processes take place. Through varying ablation studies, given coarse, noisy, and sparse conditioning inputs, we find our method to be robust and physically consistent. Part of this implementation builds upon components originally developed by Rozet, F., & Louppe, G. [paper, code] under the MIT License, which we have adapted and extended to fit our research framework.

Quickstart

  1. Install dependencies using pip or conda
pip install -r requirements.txt
  1. Run sample notebooks under notebooks/ marked with 01_ prefix. These examples are extended from [1] to benchmark against our latent approach.
  • a: Lorenz'63 system
  • b: Kolmogorov fluid

Full Experiments

In order to reproduce the results in the paper, we have to acquire the necessary data:

  1. Process the in-situ data python process_cpc.py
  2. Process the ex-situ data python process_noaa.py
  3. Process the ERA5 data from https://leap-stc.github.io/ChaosBench/quickstart.html, particularly
cd data/
wget https://huggingface.co/datasets/LEAP/ChaosBench/resolve/main/process.sh
chmod +x process.sh

./process.sh era5
./process.sh climatology
  1. Update slams/config.py field: ERA_DATADIR = <YOUR_ERA5_DIR>, for instance <PROJECT_DIR>/SLAMS/data

  2. All evaluations are summarized in a series of notebooks/ marked with 02_ prefix.

    • a: Pixel-based data assimilation
    • b: Latent-based data assimilation NO observation (only background states)
    • c: Latent-based data assimilation with +1 observation (in-situ)
    • d: Latent-based data assimilation with +2 observation (in-situ + ex-situ)
    • e: Figures and tables generation

NOTE: Training your own model is simple and is defined in train_da.py. First, define your latent model in slams/nn.py or score network in slams/score.py. Afterwards, unify both under slams/model_da.py. An example, as defined in the paper, has been provided for your reference.

Citation

If you find any of the code useful, feel free to cite these works.

@misc{qu2024deep,
      title={Deep Generative Data Assimilation in Multimodal Setting}, 
      author={Yongquan Qu and Juan Nathaniel and Shuolin Li and Pierre Gentine},
      year={2024},
      eprint={2404.06665},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

@article{rozet2024score,
  title={Score-based data assimilation},
  author={Rozet, Fran{\c{c}}ois and Louppe, Gilles},
  journal={Advances in Neural Information Processing Systems},
  volume={36},
  year={2024}
}

About

SLAMS: Score-based Latent Assimilation in Multimodal Setting


Languages

Language:Jupyter Notebook 97.6%Language:Python 2.4%