This package serves as basis for the paper: "VoMBaT: A Tool for Visualising Evaluation Measure Behaviour in High-Recall Search Tasks" by Wojciech Kusa, Aldo Lipani, Petr Knoth, Allan Hanbury
High-Recall Information Retrieval (HRIR) tasks, such as Technology-Assisted Review (TAR) used in legal eDiscovery and systematic literature reviews, focus on maximising the retrieval of relevant documents 🔎📑. Traditional evaluation measures consider precision or work saved at fixed recall levels, which can sometimes misrepresent actual system performance, especially when estimating potential savings in time and cost ⏳💰. Introducing VoMBaT – a visual analytics tool 🖥️ designed to explore the interplay between evaluation measures and varying recall levels. Our open-source tool provides insights into 18 different evaluation measures, both general and TAR-specific, letting you contrast, compare, and simulate savings in both time and money 🕵️📈️️️. Explore the metrics and their potential impacts on your HRIR tasks here.
Create and activate conda environment:
$ conda create --name tar_metrics_demo python==3.10.10
$ conda activate tar_metrics_demo
Install Python requirements:
(tar_metrics_demo)$ pip install -r requirements.txt
No additional dependencies and data are required.
Datasets' parameters are defined in data/datasets.json
file.
Start Streamlit server:
(tar_metrics_demo)$ streamlit run _🏠_Homepage.py
You can now access the app at http://localhost:8501
If you find our tool useful, please cite our paper:
@inproceedings{Kusa2023Vombat,
title = {VoMBaT: A Tool for Visualising Evaluation Measure Behaviour in High-Recall Search Tasks},
booktitle={Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval},
pages={3105--3109},
year = {2023},
doi = {https://doi.org/10.1145/3539618.3591802},
url = {https://dl.acm.org/doi/abs/10.1145/3539618.3591802},
author = {Kusa, Wojciech and Lipani, Aldo and Knoth, Petr and Hanbury, Allan}
}