Skip to content

rohanpillai20/Sentence-Similarity-based-on-Semantic-Nets-and-Corpus-Statistics

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 

Repository files navigation

Sentence Similarity based on Semantic Nets and Corpus Statistics

This is an implementation of the paper written by Yuhua Li, David McLean, Zuhair A. Bandar, James D. O’Shea, and Keeley Crockett. Link

The Sentence Similarity has been implemented as a linear combination of Semantic and Word order Similarity. Semantic and Word order Similarities are calculated from semantic and order vectors computed for each sentence with the assistance of wordnet.

Modules Required

math
os
time
sys
numpy
sklearn
nltk

  • from nltk.corupus
    • wordnet
    • brown
    • stopwords

Steps

  1. Download the 2 main programs - similarity.py and main.py.
  2. Construct the folder sub-structure as shown below:

Capture.png

  1. similarity.py has all the main functions and will be called in main.py. Compile similarity.py first to make sure there are no errors. Then call the main.py
  2. Put all the documents(text format) to be compared for similarity in the dataset sub-folder.

About

This is a full implementation of the paper written by Yuhua Li, David McLean, Zuhair A. Bandar, James D. O’Shea, and Keeley Crockett

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%