Skip to content

WojciechKusa/VoMBaT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

45 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VoMBaT: Visualisation of Evaluation Measure Behaviour in Technology Assisted Reviews

This package serves as basis for the paper: "VoMBaT: A Tool for Visualising Evaluation Measure Behaviour in High-Recall Search Tasks" by Wojciech Kusa, Aldo Lipani, Petr Knoth, Allan Hanbury

DOI:10.1145/3539618.3591802

High-Recall Information Retrieval (HRIR) tasks, such as Technology-Assisted Review (TAR) used in legal eDiscovery and systematic literature reviews, focus on maximising the retrieval of relevant documents 🔎📑. Traditional evaluation measures consider precision or work saved at fixed recall levels, which can sometimes misrepresent actual system performance, especially when estimating potential savings in time and cost ⏳💰. Introducing VoMBaT – a visual analytics tool 🖥️ designed to explore the interplay between evaluation measures and varying recall levels. Our open-source tool provides insights into 18 different evaluation measures, both general and TAR-specific, letting you contrast, compare, and simulate savings in both time and money 🕵️‍📈️️️. Explore the metrics and their potential impacts on your HRIR tasks here.

Installation

Create and activate conda environment:

$ conda create --name tar_metrics_demo python==3.10.10
$ conda activate tar_metrics_demo

Install Python requirements:

(tar_metrics_demo)$ pip install -r requirements.txt

No additional dependencies and data are required. Datasets' parameters are defined in data/datasets.json file.

Running

Start Streamlit server:

(tar_metrics_demo)$ streamlit run _🏠_Homepage.py

You can now access the app at http://localhost:8501

Citing

If you find our tool useful, please cite our paper:

@inproceedings{Kusa2023Vombat,
title = {VoMBaT: A Tool for Visualising Evaluation Measure Behaviour in High-Recall Search Tasks},
booktitle={Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval},
pages={3105--3109},
year = {2023},
doi = {https://doi.org/10.1145/3539618.3591802},
url = {https://dl.acm.org/doi/abs/10.1145/3539618.3591802},
author = {Kusa, Wojciech and Lipani, Aldo and Knoth, Petr and Hanbury, Allan}
}

Releases

No releases published

Languages