Skip to the content.
Evaluation
- Evaluation measures / scorer
- The performance of the predictors will be evaluated in terms of the RAE (relative absolute error) and AE (absolute error) measure; only RAE will be used for the final ranking, though. The exact evaluation script that will be used to evaluate the participants submissions is available here. Check Fabrizio Sebastiani:
Evaluation measures for quantification: An axiomatic approach. Information Retrieval Journal 23(3): 255-288 (2020) for a thorough discussion of RAE, AE, and their suitability to evaluating quantification systems.
- The test set will consist of a number of test samples (i.e., sets of documents), some of them characterized by prevalence values very different from the ones that chatacterize the training set; this is done in order to test the robustness of the quantifier to predict class prevalences very different from those it has been trained on. Check paper Andrea Esuli, Alejandro Moreo, and Fabrizio Sebastiani. LeQua@CLEF2022: Learning to Quantify. Proceedings of the 44th European Conference on Information Retrieval (ECIR 2022), Stavanger, NO. for a description of how the test samples have been generated
- A scorer that reflects how the evaluation is carried out will be released to participants by December 1, 2021.
- Tools and baselines
- Required format for submissions