An implementation of the INST information retrieval evaluation measure in the style of trec_eval.
INST is an evaluation measure outlined in:
A. Moffat, P. Bailey, F. Scholer, P. Thomas. INST: An Adaptive Metric for Information Retrieval Evaluation. ADCS'15 Proceedings of the 20th Australasian Document Computing Symposium. Sydney, Australia, December 2015.
We thank Moffat and Bailey for providing an initial implementation of INST from which this code was developed.
If you do use our inst_eval implementation then please include citiations to both the Moffat et al. paper and to our paper:
B. Koopman and G. Zuccon. A test collection for matching patient trials. In Proceedings of the 39th annual international ACM SIGIR conference on research and development in information retrieval, Pisa, July 2016.
Setting up inst_eval.py
inst_eval.py is run as a standalone program. It requires Python 2.7 to be installed.
For usage information run:
This will print the following:
usage: inst_eval.py [-h] [-c] [-n EVAL_DEPTH] [-q] (-tpq T_PER_QUERY | -T OVER_WRITE_T) trec_qrel_file trec_results_file Implementation of the INST evaluation measure from 'INST: An Adaptive Metric for Information Retrieval Evaluation', ACDS2015. positional arguments: trec_qrel_file TREC style qrel file. trec_results_file TREC style results file. optional arguments: -h, --help show this help message and exit -c, --complete_qrel_queries Same as -c in trec_eval: Average over the complete set of queries in the relevance judgements instead of the queries in the intersection of relevance judgements and results. Missing queries will contribute a value of 0 to all evaluation measures -n EVAL_DEPTH, --eval_depth EVAL_DEPTH EVAL_DEPTH: Max depth to evaluate at -q, --per_query Print out per query evaluation result -tpq T_PER_QUERY, --T_per_query T_PER_QUERY Tab separated file indicating value of T for each query: QueryId<tab>T -T OVER_WRITE_T, --over_write_T OVER_WRITE_T Set all T values to supplied constant.