Recent studies suggest that changes to global climate as have been seen at the Earth's land and ocean surface are also making their way into the deep ocean, which is the largest active storage system for heat and carbon available on the timescale of a human lifetime. Historical measurements of subsurface ocean temperature are essential to the scientific research investigating the changes in the amount of heat stored in the ocean and also to other climate research activities such as combining observations with numerical models to provide estimates of the global ocean's and Earth's climate state in the past and predictions for the future. Unfortunately, as with all observations, these measurements contain errors and biases that must be identified to prevent a negative impact on the applications and investigations that rely on them. Various groups from around the world have developed quality control tests to perform this important task. However, this has led to duplication of effort, code that is not easily available to other researchers and the introduction of climate model differences solely due to the varying performance of these software systems whose nuances relative to one another are poorly known.
Recently, an international team of researchers has decided to work together to break down the barriers between the various groups and countries through the formation of the IQuOD (International Quality Controlled Dataset) initiative. One of the key aims is to intercompare the performance of the various automatic quality control tests that are presently being run to determine a best performing set. This work has started. However, it currently involves individuals running test datasets through their own systems and is being confounded by complications associated with the differences in the file formats and systems that are in use in the various labs and countries.
The IQuOD proposal is to set up an open quality control benchmarking system. Work will begin by implementing a battery of simple tests to run on some test data, and producing summary statistics and visualizations of the results. Later goals include helping researchers either wrap their existing C, Fortran and MATLAB test functions in Python for use in this test suite, or re-implementing those tests in native Python.
Dependencies & Setup:
Tested on Ubuntu 16.04
To clone this project and set it up, make sure
git is installed, then:
$ git clone https://github.com/IQuOD/AutoQC $ cd AutoQC $ source install.sh
To run AutoQC in a containerized environment, make sure
docker is installed, then:
$ docker image run -it -v /my/data/directory:/rawdata iquod/autoqc:ubuntu-16.04 bash
/my/daya/directory on your machine will be available at
/rawdata inside the container, and vice versa. Use this to add raw WOD-ASCII data to your container, and add multiple
-v origin:destination paths to include multiple directories in the same way.
You may also want to
git pull origin master inside the
/AutoQC directory inside your container, to fetch the latest version of the project.
AutoQC runs in three steps: database construction, qc running, and result summarization.
python build-db.py filename tablename
filename is the name of a WOD-ascii file to read profiles from, and
tablename is the name of a postgres table to write the results to;
tablename will be created if it doesn't
exist, or appended to if it does.
tablename will have the following columns:
||the raw WOD-ASCII text originally found in the input file|
||whether any temperature qc levels were flagged at 3 or greater|
||unique profile serial number|
||timestamp month, integers [1,12]|
||timestamp day, integers [1,31]|
||timestamp walltime, real [0,24)|
||probe index, per WOD specifications|
Additionally, there is a column in the table for the qc results of every test found in the
/qctests directory; these columns are filled in in the next step.
python AutoQC.py tablename nProcessors
tablename is the postgres table to pull profiles from (probably the same as
tablename in the last step), and
nProcessors is how many processors you'd like to parallelize over.
python summarize-results.py tablename
tablename is the postgres table used in the previous steps. A summary of true flags, true passes, false positives and false negatives is generated for each test.
Each quality control test must be written as its own file in
/qctests, of the form
def test(p, parameters), where
p is a profile object; each test returns a bool, where
True indicates the test has failed.
parameters is a dictionary for conveniently persisting static parameters and sharing them between threads; if your test has a great deal of parameters to load before it runs, include alongside its definition a
loadParmaeters(dict) method, which writes those
parameters to keys of your choosing on the dictionary passed in as an argument to
loadParameters. That dictionary will subsequently be passed into every qc test as the
parameters argument. Calling this
loadParameters function is done automatically by the qc framework;
it is enough for you to just write it, and the parameters you want will be available in your qc test on the keys you defined on the
To run the code tests:
pip install nose nosetests tests/*.py
Profile Objects Specification
See wodpy package for more information on the WodProfile class, a decoding helper for the WOD ASCII format.
##Contributing Quality control checks waiting to be implemented are listed in the Issues. If you would like to work on coding up a check, please assign yourself to the issue to avoid others duplicating the effort. If you have an idea for a new QC check, please open an issue and let us know, so we can help get you started on the right track.