Providers¶
accuracy
¶
Provider for the accuracy metric
compat
¶
- Version of the compatibility measure desribed in
- @article{10.1145/3451161,
author = {Clarke, Charles L. A. and Vtyurina, Alexandra and Smucker, Mark D.}, title = {Assessing Top-k Preferences}, journal = {ACM Transactions on Information Systems}, volume = {39}, number = {3}, articleno = {33}, numpages = {21}, year = {2021}, url = {https://doi.org/10.1145/3451161},
}
Supported Measures:
Compat(p=ANY, normalize=ANY)
cwl_eval
¶
cwl_eval, providing C/W/L (“cool”) framework measures.
@inproceedings{azzopardi2019cwl,
author = {Azzopardi, Leif and Thomas, Paul and Moffat, Alistair},
title = {cwl\_eval: An Evaluation Tool for Information Retrieval},
booktitle = {SIGIR},
year = {2019}
}
Supported Measures:
P(rel=ANY)@ANY
RR(rel=ANY)@NOT_PROVIDED
AP(rel=ANY)@NOT_PROVIDED
RBP(rel=REQUIRED, p=ANY)@NOT_PROVIDED
BPM(T=ANY, min_rel=ANY, max_rel=REQUIRED)@ANY
SDCG(dcg='log2', min_rel=ANY, max_rel=REQUIRED)@REQUIRED
NERR8(min_rel=ANY, max_rel=REQUIRED)@REQUIRED
NERR9(min_rel=ANY, max_rel=REQUIRED)@REQUIRED
NERR10(p=ANY, min_rel=ANY, max_rel=REQUIRED)
NERR11(T=ANY, min_rel=ANY, max_rel=REQUIRED)
INST(T=ANY, min_rel=ANY, max_rel=REQUIRED)
INSQ(T=ANY, min_rel=ANY, max_rel=REQUIRED)
gdeval
¶
gdeval
Supported Measures:
nDCG(dcg='exp-log2')@REQUIRED
ERR@REQUIRED
judged
¶
python implementation of judgment rate
Adapted from OpenNIR’s implementation: https://github.com/Georgetown-IR-Lab/OpenNIR/blob/master/onir/metrics/judged.py
Supported Measures:
Judged@ANY
msmarco
¶
MS MARCO’s implementation of RR
Supported Measures:
RR(rel=ANY)@ANY
pyndeval
¶
pyndeval
Supported Measures:
ERR_IA(rel=ANY, judged_only=ANY)@ANY
nERR_IA(rel=ANY, judged_only=ANY)@ANY
alpha_DCG(alpha=ANY, rel=ANY, judged_only=ANY)@ANY
alpha_nDCG(alpha=ANY, rel=ANY, judged_only=ANY)@ANY
NRBP(alpha=ANY, beta=ANY, rel=ANY)
nNRBP(alpha=ANY, beta=ANY, rel=ANY)
AP_IA(rel=ANY, judged_only=ANY)
P_IA(rel=ANY, judged_only=ANY)@ANY
StRecall(rel=ANY)@ANY
pytrec_eval
¶
pytrec_eval
@inproceedings{VanGysel2018pytreceval,
title={Pytrec\_eval: An Extremely Fast Python Interface to trec\_eval},
author={Van Gysel, Christophe and de Rijke, Maarten},
publisher={ACM},
booktitle={SIGIR},
year={2018},
}
Supported Measures:
P(rel=ANY)@ANY
RR(rel=ANY)@NOT_PROVIDED
Rprec(rel=ANY)
AP(rel=ANY)@ANY
nDCG(dcg='log2')@ANY
R@ANY
Bpref(rel=ANY)
NumRet(rel=ANY)
NumQ
NumRel(rel=1)
SetAP(rel=ANY)
SetF(rel=ANY, beta=ANY)
SetP(rel=ANY, relative=ANY)
SetR(rel=ANY)
Success(rel=ANY)@ANY
IPrec@ANY
infAP(rel=ANY)
ranx
¶
@misc{ranx2021,
title = {ranx: A Blazing-Fast Python Library for Ranking Evaluation and Comparison},
author = {Bassani, Elias},
year = {2021},
publisher = {GitHub},
howpublished = {\url{https://github.com/AmenRa/ranx}},
}
Supported Measures:
P(rel=ANY)@ANY
SetP(rel=ANY)
RR(rel=ANY)@NOT_PROVIDED
Rprec(rel=ANY)
AP(rel=ANY)@ANY
nDCG(dcg=('log2', 'exp-log2'))@ANY
R@ANY
SetR(rel=ANY)
NumRet(rel=REQUIRED)
Success(rel=ANY)@REQUIRED
trectools
¶
trectools
@inproceedings{palotti2019,
author = {Palotti, Joao and Scells, Harrisen and Zuccon, Guido},
title = {TrecTools: an open-source Python library for Information Retrieval practitioners involved in TREC-like campaigns},
series = {SIGIR'19},
year = {2019},
location = {Paris, France},
publisher = {ACM}
}
Supported Measures:
P(rel=1)@ANY
RR(rel=1)@NOT_PROVIDED
Rprec(rel=1)
AP(rel=1)@ANY
nDCG(dcg=ANY)@ANY
Bpref(rel=1)
RBP(p=ANY, rel=ANY)@ANY