Skip to content

lwezenbe/correctionlib

 
 

Repository files navigation

correctionlib

Actions Status Documentation Status Code style: black

PyPI version PyPI platforms

GitHub Discussion

Introduction

The purpose of this library is to provide a well-structured JSON data format for a wide variety of ad-hoc correction factors encountered in a typical HEP analysis and a companion evaluation tool suitable for use in C++ and python programs. Here we restrict our definition of correction factors to a class of functions with scalar inputs that produce a scalar output.

In python, the function signature is:

from typing import Union

def f(*args: Union[str,int,float]) -> float:
    return ...

In C++, the evaluator implements this currently as:

double Correction::evaluate(const std::vector<std::variant<int, double, std::string>>& values) const;

The supported function classes include:

  • multi-dimensional binned lookups;
  • binned lookups pointing to multi-argument formulas with a restricted math function set (exp, sqrt, etc.);
  • categorical (string or integer enumeration) maps;
  • input transforms (updating one input value in place); and
  • compositions of the above.

Each function type is represented by a "node" in a call graph and holds all of its parameters in a JSON structure, described by the JSON schema. Possible future extension nodes might include weigted sums (which, when composed with the others, could represent a BDT) and perhaps simple MLPs.

The tool should provide:

  • standardized, versioned JSON schemas;
  • forward-porting tools (to migrate data written in older schema versions); and
  • a well-optimized C++ evaluator and python bindings (with numpy vectorization support).

This tool will definitely not provide:

  • support for TLorentzVector or other object-type inputs (such tools should be written as a higher-level tool depending on this library as a low-level tool)

Formula support currently includes a mostly-complete subset of the ROOT library TFormula class, and is implemented in a threadsafe standalone manner. The parsing grammar is formally defined and parsed through the use of a header-only PEG parser library. The supported features mirror CMSSW's reco::formulaEvaluator and fully passes the test suite for that utility with the purposeful exception of the TMath:: namespace. The python bindings may be able to call into numexpr, though, due to the tree-like structure of the corrections, it may prove difficult to exploit vectorization at levels other than the entrypoint.

Installation

The build process is based on setuptools, with CMake (through scikit-build) for the C++ evaluator and its python bindings module. Builds have been tested in Windows, OS X, and Linux, and python bindings can be compiled against both python2 and python3, as well as from within a CMSSW environment. The python bindings are distributed as a pip-installable package.

To build in an environment that has python 3, you can simply

pip install correctionlib

(possibly with --user, or in a virtualenv, etc.) Note that CMSSW 11_2_X and above has ROOT accessible from python 3.

The C++ evaluator is part of the python package. If you are also using CMake you can depend on it by passing -Dcorrectionlib_DIR=$(python -c 'import pkg_resources; print(pkg_resources.resource_filename("correctionlib", "cmake"))'). The header and shared library can similarly be found as

import pkg_resources
pkg_resources.resource_filename("correctionlib", "include/correction.h")
pkg_resources.resource_filename("correctionlib", "lib/libcorrectionlib.so")

In environments where no recent CMake is available, or if you want to build against python2 (see below), you can build the C++ evaluator via:

git clone --recursive git@github.com:nsmith-/correctionlib.git
cd correctionlib
make
# demo C++ binding, main function at src/demo.cc
gunzip data/examples.json.gz
./demo data/examples.json

Eventually a correction-config utility will be added to retrieve the header and linking flags.

To compile with python2 support, consider using python 3 :) If you considered that and still want to use python2, follow the C++ build instructions and then call make PYTHON=python2 correctionlib to compile. Inside CMSSW you should use make PYTHON=python correctionlib assuming python is the name of the scram tool you intend to link against. This will output a correctionlib directory that acts as a python package, and can be moved where needed. This package will only provide the correctionlib._core evaluator module, as the schema tools and high-level bindings are python3-only.

Creating new corrections

The correctionlib.schemav2 module provides a helpful framework for defining correction objects and correctionlib.convert includes select conversion routines for common types. Nodes can be type-checked as they are constructed using the parse_obj class method or by directly constructing them using keyword arguments. Some examples can be found in data/conversion.py. The tests/ directory may also be helpful.

Developing

See CONTRIBUTING.md

About

A generic correction library

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 64.1%
  • C++ 32.0%
  • CMake 2.9%
  • Makefile 1.0%