Skip to content

vrunm/Text-Summarization-News-Articles

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

50 Commits
 
 
 
 
 
 

Repository files navigation

Text-Summarization-Multi-News

Data:

Multi-News, consists of news articles and human-written summaries of these articles from the site newser.com. Each summary is professionally written by editors and includes links to the original articles cited.

There are two features:

document: Text of News Articles
summary: News Summary.

Experiments:

BART:

The BART model for text summarization was trained with Adam optimizer having learning rate 2e-5 for 6 epochs yielded a ROUGUE-L score of 38 and F1(weighted) of 0.84.

DistilBART:

The DistilBART model for text summarization was trained with Adam optimizer having learning rate 2e-5 for 6 epochs yielded a ROUGUE-L score of 42 and F1(weighted) of 0.86.

Evaluation Metrics:

The ROUGE metric was used for evaluation:

ROUGE-N measures the number of matching ‘n-grams’ between our model-generated text and a ‘reference’.

An n-gram is simply a grouping of tokens/words. A unigram (1-gram) would consist of a single word. A bigram (2-gram) consists of two consecutive words:

For ROUGE-1 we would be measuring the match-rate of unigrams between our model output and reference.
ROUGE-2 and ROUGE-3 would use bigrams and trigrams respectively.

Model Epochs ROUGUE-L F1 Score(Weighted)
BART 6 38 0.84
DistilBART 6 42 0.86
Optimizer Learning Rate $\gamma$ Momentum $\eta$ Alpha $\alpha$ Beta1 $\beta_1$ Beta2 $\beta_2$ Epsilon $\epsilon$
AdamW 5e-5 0.01 0.9 0.9 0.999 1e-5
RMSprop 0.01 0.01 0.99 - - -
NAG 5e-5 - - - -
SGD(Momentum) 5e-5 0.001 - - - -
SGD 0.01 - - - -

The Finetuning of the optimizers was based on: On Empirical Comparisons of Optimizers for Deep Learning
Comparing the Training loss of all optimizers

Comparing the Validation loss of all optimizers

The rate of convergence of the Adam optimizer is the fastest.

We can conclude the order of convergence of the optimizers: AdamW > RMSprop > NAG > SGD (Momentum) > SGD

About

Built a summarization model for news articles.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages