Skip to content

grey-nearing/da_efficiency

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Code repository for the paper: The Efficiency of Data Assimilation
https://agupubs.onlinelibrary.wiley.com/doi/abs/10.1029/2017WR020991

Welcome to the REAMDE file!  

How To Run:

1) Compile Noah-MP model:
	1a) With an ifort compiler, use the makefile.discover as a template. On NCCS Discover, this can be accessed by sourcing the <environment_setup.sh> file. Note that this purges all modules and changes $PATH.
        1a) With a gfortran compiler, use the makefile as a template.
	1b) In the <model_code> subdirectory run 'make clean' and 'make'.

2) Make ensemble forcing data:
	2a) Navigate to the <data/forcing> subdirectory
	2b) Run the <pert_force.m> MatLab script.

3) CDF-match the LPRM observations:
	3a) Navigate to the <data/lprm> subdirectory.
	3b) Run the <cdf_matching.m> matlab script.

4) Run the open-loop simulations:
	4a) Naviage to the <open> subdirectory.
	4b) Create runtime subdirectories by running <setup_dirs.sh> script.
	4c) Edit the hard-paths and header information in the <job.slurm> SLURM/PODS script [*].
	4d) Run the open-loop simulations by submitting the <job.slurm> SLURM file.

5) Run the EnKF-loop simulations:
	5a) Navigate to the <enkf> subdirectory.
	5b) Create runtime subdirectories by running <setup_dirs.sh> script.
	5c) Edit the hard-paths and header information in the <job.slurm> SLURM/PODS script [*].
	5d) Run the open-loop simulations by submitting the <job.slurm> SLURM file.

6) Run the various analysis MatLab scripts in the <analysis> subdirectory to create the various figures in the paper.
	6a) <make_scan_map> makes Figure 3 - showing locations of SCAN evaluation data stations.
	6b) <main_convergence> makes Figure 4 - showing convergence of the IT metrics as a function of sample size.
	6c) <main_efficiency> makes Figure 5 & Table 3 - showing main efficiency metrics.
	6d) <main_decomposition> makes Figure 6 & Table 4 - showing decomposition metrics. 
	6e) <sig_vs_info> makes Figure 7 - comparing Gaussian vs. Information-Theory statistics.

[*] On computers other than NCCS Discover, the user will likely have to set up their own job scheduling routines. This one is specifically designed for SLURM with PODS, and uses modules availalbe on Discover.


About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages