Python implementation of the INST evaluation measure from Moffat et al, ADCS 2015.
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Failed to load latest commit information.

An implementation of the INST information retrieval evaluation measure in the style of trec_eval.

About INST

INST is an evaluation measure outlined in:

A. Moffat, P. Bailey, F. Scholer, P. Thomas. INST: An Adaptive Metric for Information Retrieval Evaluation. ADCS'15 Proceedings of the 20th Australasian Document Computing Symposium. Sydney, Australia, December 2015.

We thank Moffat and Bailey for providing an initial implementation of INST from which this code was developed.

If you do use our inst_eval implementation then please include citiations to both the Moffat et al. paper and to our paper:

B. Koopman and G. Zuccon. A test collection for matching patient trials. In Proceedings of the 39th annual international ACM SIGIR conference on research and development in information retrieval, Pisa, July 2016.

Setting up is run as a standalone program. It requires Python 2.7 to be installed.


For usage information run:

./ -h

This will print the following:

usage: [-h] -n EVAL_DEPTH [-T OVER_WRITE_T]
                    trec_qrel_file trec_results_file T_per_query

Implementation of the INST evaluation measure from 'INST: An Adaptive Metric
for Information Retrieval Evaluation', ACDS2015.

positional arguments:
  trec_qrel_file        TREC style qrel file.
  trec_results_file     TREC style results file.
  T_per_query           Tab separated file indicating value of T for each
                        query: QueryId<tab>T

optional arguments:
  -h, --help            show this help message and exit
  -n EVAL_DEPTH, --eval_depth EVAL_DEPTH
                        Max depth to evaluate at.
  -T OVER_WRITE_T, --over_write_T OVER_WRITE_T
                        Set all T values to supplied constant.