End-to-end Snakemake workflow to run Monte Carlo simulations of signal and background signatures in the LEGEND experiment and produce probability-density functions (pdfs). Configuration metadata (e.g. rules for generating simulation macros or post-processing settings) is stored at legend-simflow-config.
- Simulations are labeled by an unique identifier (e.g.
l200a-hpge-bulk-2vbb
), often referred assimid
(simID). The identifier is defined in legend-simflow-config throughsimconfig.yaml
files in tier directoriesstp
andver
. - Each simulation is defined by a template macro (also stored as metadata) and by a set of rules
(in
simconfig.yaml
) needed to generate the actual macros (template variable substitutions, number of primaries, number of jobs, etc). - The production is organized in tiers. The state of a simulation in a certain tier is labeled
as
<tier>.<simid>
. Snakemake understands this syntax. - The generated pdfs refer to a user-defined selection of LEGEND data taking runs. Such a list of runs is specified through the configuration file.
- The production can be restricted to a subset of simulations by passing a list of identifiers to Snakemake.
- Macro files are generated and writted to disk according to rules defined in the metadata. These macros are in one-to-one correspondence with simulation jobs in the next tiers
- Tier
ver
building: run simulations that generate Monte Carlo event vertices needed to some simulations in the next tier. Simulations that do not need a special event vertices will directly start from tierraw
. - Tier
raw
building: run full event simulations. - Tier
hit
building: run the first (hit-oriented) step of simulation post-processing. Here, "hit" stands for Geant4 "step". In this tier, step-wise operations like optical map application or step clustering are typically applied. - Tier
evt
building: multiple operations are performed in order to build actual events and incorporate information about the data taking runs for which the user wants to build pdfs:- Partition the
hit
event statistics into fractions corresponding to the actual total livetime fraction spanned by each selected run. This information is extracted fromlegend-metadata/dataprod/runinfo.json
- Apply HPGe energy resolution functions for each selected run found in the data production auto-generated metadata
- Apply HPGe status flags (available in
legend-metadata/hardware/config
)
- Partition the
- Tier
pdf
building: summarizeevt
-tier output into histograms (the pdfs).
The config.yaml
file in the production directory allows to customize the workflow
in great detail. Here's a basic description of its fields:
experiment
: labels the experiment to be simulated. The same name is used in the metadata to label the corresponding configuration files.simlist
: list of simulation identifiers (see below) to be processed by Snakemake. Can be a list of strings or a path to a text file. If*
orall
, will process all simulations defined in the metadata.runlist
: list of LEGEND data taking runs to build pdfs for, in the standard format<experiment>-<period>-<run>-<type>
(e.g.l200-p03-r000-phy
)benchmark
: section used to configure a benchmarking run:enabled
: boolean flag to enable/disable the featuren_primaries
: number of primary events to be simulated in the lower tiersver
andraw
.
paths
: customize paths to input or output files.execenv
: defines the software environment (container) where all jobs should be executed (see below).
Tip
all these configuration parameters can be overridden at runtime through Snakemake's --config
option.
Run a production by using one of the provided site-specific profiles (recommended):
> cd <path-to-cycle-directory>
> snakemake --workflow-profile workflow/profiles/<profile-name>
If no system-specific profiles are provided, the --workflow-profile
option can be omitted.
Snakemake will use the default
profile.
> snakemake
The --config
command line option is very useful to override configuration values.
It can be used, for example, to restrict the production to a subset of simulations:
> snakemake --config simlist="mylist.txt" [...]
where mylist.txt
is a text file in the format:
raw.l200a-fibers-Ra224-to-Pb208
hit.l200a-hpge-bulk-2vbb
...
One can even just directly pass a comma-separated list:
> snakemake --config simlist="raw.l200a-fibers-Ra224-to-Pb208,hit.l200a-hpge-bulk-2vbb"
Once the production is over, the print_stats
rule can be used to display a table with runtime statistics:
> snakemake -q all print_stats
wall time [s] wall time [s]
simid (cumulative) jobs (per job) primaries
----- ------------- ---- ------------- ---------
hit.l200a-fibers-Ra224-to-Pb208 83:20:00 100 0:50:00 1.00E+08
raw.l200a-fibers-Ra224-to-Pb208 58:20:35 100 0:35:00 1.00E+08
raw.l200a-fibers-Rn222-to-Po214 33:20:00 100 0:20:00 1.00E+08
... ... ... ... ...
The inspect_simjob_logs
rule allows to inspect the simulation log files in search for warnings:
> snakemake inspect_simjob_logs
This can generate a lot of output, consider piping it to a file.
Find some useful Snakemake command-line options at the bottom of this page.
This workflow implements the possibility to run special "benchmarking" runs in order to evaluate the speed of simulations, for tuning the number of events to simulate for each simulation run.
- Create a new, dedicated production cycle (see above)
- Enable benchmarking in the configuration file and customize further settings
- Start the production as usual.
Snakemake will spawn a single job (with the number of primary events specified in the
configuration file) for each simulation. Once the production is over, the results can
be summarized via the print_benchmark_stats
rule:
> snakemake -q all print_benchmark_stats
simid CPU time [ms/ev] evts / 1h jobs (1h) / 10^8 evts
----- ---------------- --------- ---------------------
raw.l200a-birds-nest-K40 (13s) 2.79 1288475 77
raw.l200a-birds-nest-Ra224-to-Pb208 (191s) 38.33 93916 1064
raw.l200a-fiber-support-copper-Co60 (223s) 44.69 80558 1241
... ... ... ...
Note
The CPU time is a good measure of the actual simulation time, since other tasks (e.g. application loading) are typically not CPU intensive.
Start the production on the interactive node (the default profile works fine):
snakemake
Start the production on the batch nodes (via SLURM):
snakemake --workflow-profile workflow/profiles/nersc-batch
usage: snakemake [OPTIONS] -- [TARGET ...]
--dry-run, --dryrun, -n
Do not execute anything, and display what would be done. If you have a very large workflow,
use --dry-run --quiet to just print a summary of the DAG of jobs.
--profile PROFILE Name of profile to use for configuring Snakemake.
--cores [N], -c [N] Use at most N CPU cores/jobs in parallel. If N is omitted or 'all', the limit is set to the
number of available CPU cores.
--config [KEY=VALUE ...], -C [KEY=VALUE ...]
Set or overwrite values in the workflow config object.
--configfile FILE [FILE ...], --configfiles FILE [FILE ...]
Specify or overwrite the config file of the workflow.
--forceall, -F Force the execution of the selected (or the first) rule and all rules it is dependent on
regardless of already created output.
--list, -l Show available rules in given Snakefile. (default: False)
--list-target-rules, --lt
Show available target rules in given Snakefile.
--summary, -S Print a summary of all files created by the workflow.
--printshellcmds, -p Print out the shell commands that will be executed. (default: False)
--quiet [{progress,rules,all} ...], -q [{progress,rules,all} ...]
Do not output certain information. If used without arguments, do not output any progress or
rule information. Defining 'all' results in no information being printed at all.