Skip to content

bigabig/faithfulness

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Faithfulness 😇

An easy-to-use library to evaluate faithfulness (factual correctness) of abstractive summaries. Faithfulness is computed by comparing a summary with its original source document.

This library includes multiple faithfulness metrics based on:

  • BERTScore
  • Entailment
  • Question Generation & Question Answering framework (QGQA)
  • Named Entity Recognition (NER)
  • Open Information Extraction (OpenIE)
  • Semantic Role Labeling (SRL)
  • Sentence Similarity (SentSim)

Installation ⚙️

  1. $ conda create -n my_project python=3.8 This creates a new virtual environment for your project with conda. You can activate it with $ conda activate my_project.
  2. $ conda install pytorch==1.7.1 torchvision==0.8.2 torchaudio==0.7.2 cudatoolkit=10.1 -c pytorch Please install PyTorch by following the instructions here. Make sure to install the CUDA variant that matches the CUDA version of your GPU.
  3. $ pip install faithfulness This installs the faithfulness library and it's dependencies. Read more about the dependencies below.

All faithfulness metrics are model-based. Some models have to be installed manually:

  • Download the SRL model here and save it in your project. e.g. /models/srl_model.tar.gz
  • Download a spacy model: $ python -m spacy download en_core_web_sm
  • Download CoreNLP: import stanza && stanza.install_corenlp()

Usage 🔥

from faithfulness.QGQA import QGQA

qgqa = QGQA()
summary = "Lorem ipsum dolor sit amet"
source = "Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam ..."
result: QGQAResult = qgqa.score(summary, source)
print(f"Faithfulness: {result["f1"]}")

More examples can be found here 💯.

Evaluation 📊

We evaluated all faithfulness metrics by correlating them with human judgements on the XSUM dataset (link). You will soon be able to read more about the evaluation in our paper. (Master's thesis)

Method Pearson (r) Spearman (p)
🥇 BERTScore 0.501 0.486
🥈 Entailment 0.366 0.422
🥉 SentSim 0.392 0.389
SRL 0.393 0.377
NER 0.252 0.259
QGQA 0.228 0.258
OpenIE 0.169 0.185

Reproduce results & evaluate custom dataset

You can download the preprocessed XSUM dataset here and the preprocessed Summeval dataset here to reproduce the above results. To evaluate the faithfulness metrics on other datasets, we recommend using the provided Experimentor class. For this, your dataset has to be in the following JSON format:

prepared_dataset.json

[{
    "summary": "the summary text..."
    "source": "the source text..."
    "summary_sentences": ["summary sentence 1", "summary sentence 2", ...] 
    "source_sentences": ["source sentence 1", "source sentence 2", ...] 
    "faithfulness": 0.0 - 1.0
}, ...]

You can convert .csv files to the required format using Experimentor.prepare_dataset(). This function uses the spacy model "en_core_web_lg" to perform sentence splitting and requires a csv file with summary, source and faithfulness keys:

dataset.csv:
summary,source,faithfulness
"summary text","source text",0.9

Experimentor.prepare_dataset(Path("dataset.csv"))

You can now use the experimentor:

output_path=Path("./experiments/dataset/qgqa"
faithfulness_metric = QGQA(metric=BERTScore, save_path=output_path, batch_mode=True)
Experimentor(data_path=Path("./prepared_dataset.json"),
             out_path=output_path),
             metric=faithfulness_metric,
             experiment_name="qgqa_bertscore").experiment()

In the above example, correlations are written to /experiments/dataset/qgqa/qgqa_bertscore.csv

Dependencies 🔗

By running $ pip install faithfulness you will install this library as well as the following dependencies:

Troubleshooting 🛠

There are currently problems when installing allennlp and jsonnet. If you encounter "Building wheel for jsonnet (setup.py) ... error_" during the installation please try:

apt-get install make 
apt-get install g++ 

or install jsonnet before installing this library

conda install -c conda-forge jsonnet
pip install faithfulness

About

A library to evaluate factual correctness of abstractive summaries.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages