You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Broad releases new data every quarter that is processed by this app.
Currently updating for this data is a multistep process:
@hirscheylab performs some data validation, then creates a PR that updates release name, URLs and methods details. Example PR: 20Q1 update #57
After the above PR is merged a docker image is automatically built. Wait for the docker image to build.
On our HPC cluster I generate the data and upload the results to a DukeDS project. This process uses a singularity image created from the above docker image and a clone of this repo.
I would like to simplify this process or at least have these details recorded so I don't forget them.
Notes based on steps above
Step 3 - Generate data - Docker Image
The docker image is used to supply the r libraries used by data generation. I currently also clone the repo and use them in combination. I am wondering if I could use just the docker/singularity image.
Step 3 - Generate data - directories to create
I need to manually create the following directories after cloning the repo: logs, singularity/images, and data. I'm not sure why the Makefile doesn't create the data
directory.
Step 3 - Generate data - sbatch commands
We have some notes here: https://github.com/hirscheylab/ddh#singularity
Basically I setup a config file, run sbatch build-slurm.sh wait for it to finish successfully then run sbatch upload-slurm.
Step 4 - Update website - Update list of files to download
To download data in the openshift app requires installing and configuring the openshift oc command.
To rerun the job to download data in the openshift app usually requires deleting the previous job:
oc delete job download-ddh-data
Then creating/running the job to download data:
oc create -f DownloadJob.yaml
On the job finishes I redeploy the website using the okd application console: depmap -> Applications -> ddh-shiny-app -> Click Deploy.
Broad releases new data every quarter that is processed by this app.
Currently updating for this data is a multistep process:
openshift/file-list.json
.I would like to simplify this process or at least have these details recorded so I don't forget them.
Notes based on steps above
Step 3 - Generate data - Docker Image
The docker image is used to supply the r libraries used by data generation. I currently also clone the repo and use them in combination. I am wondering if I could use just the docker/singularity image.
Step 3 - Generate data - directories to create
I need to manually create the following directories after cloning the repo:
logs
,singularity/images
, anddata
. I'm not sure why the Makefile doesn't create thedata
directory.
Step 3 - Generate data - sbatch commands
We have some notes here: https://github.com/hirscheylab/ddh#singularity
Basically I setup a config file, run
sbatch build-slurm.sh
wait for it to finish successfully then runsbatch upload-slurm
.Step 4 - Update website - Update list of files to download
To update the list of files to download into openshift we have openshift/make-file-list.py. This script creates a file for all files in the DukeDS project and not just the current release so I manually remove the older files from this list.
Step 4 - Update website - Download
To download data in the openshift app requires installing and configuring the openshift oc command.
To rerun the job to download data in the openshift app usually requires deleting the previous job:
Then creating/running the job to download data:
On the job finishes I redeploy the website using the okd application console: depmap -> Applications -> ddh-shiny-app -> Click Deploy.
FYI: @dleehr
The text was updated successfully, but these errors were encountered: