Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Differentiable analysis: from scratch vs existing methods #4

Open
phinate opened this issue May 11, 2020 · 1 comment
Open

Differentiable analysis: from scratch vs existing methods #4

phinate opened this issue May 11, 2020 · 1 comment

Comments

@phinate
Copy link
Member

phinate commented May 11, 2020

An interesting point from the kickoff meeting was that there are broadly two directions that any new tools should try to consider:

  • Creating a differentiable workflow from scratch
  • Interfacing with existing approaches

It would be very constructive to highlight the considerations one should make for each case :)

E.g. Kyle Cranmer pointed out that to be truly ‘optimal’, invariant of approach, one should learn the true likelihood function instead of objectives like those targeted by INFERNO (inverse fisher info) or neos (p-value from hyp test).

@pablodecm
Copy link

pablodecm commented May 14, 2020

I agree that learning the fully parameterized likelihood, likelihood ratio or score (both with nuisance and interest parameters) are truly optimal in a more strict sense. That said, training a well-performing parametrized model like that can be challenging when many parameters are considered, and then in the case of likelihood and likelihood-ratio it needs to be cross checked and calibrated that can also be difficult. The SALLY and SALLINO are somehow more similar to INFERNO and neos in that a summary statistic can be constructed from their output and used with the same tools that are currently used.

That said I think this are not two different directions perse, some of the most efficient methods in MadMiner require gradients with respect to statistical model parameters, so if we build a set of tools that has the abstractions tthat allow to construct a differentiable functional version of sets of events and their weight (see my answer in #7 (comment) ) as a function of nuisance parameters and parameters of interest all the approaches could benefit.

I think the idea of having a set of modules, abstractions and examples for building that first part could be useful for all approaches. Then you could plug and play different training modules and objective functions, which could be also implemented in a common framework.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants