Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Infersent #71

Open
Radhikadua123 opened this issue Jun 13, 2019 · 1 comment
Open

Infersent #71

Radhikadua123 opened this issue Jun 13, 2019 · 1 comment

Comments

@Radhikadua123
Copy link

Hi,

I am trying to obtain the semantic similarity between the generated and the ground truth sentence.

I used all these metrics to evaluate the generated sentences (validation dataset):
BLEU 1 | 0.128031
BLEU 2 | 0.056153
BLEU 3 | 0.029837
BLEU 4 | 0.013649
METEOR | 0.305482
ROUGE_L | 0.148652
CIDEr | 0.069519
SkipThought cosine similarity | 0.765784
Embedding Average cosine similarity | 0.973187
Vector Extrema cosine similarity | 0.683888
Greedy Matching score | 0.94496

Some of these metrics indicates the sentences to be quite similar and some shows sentences to be different. Can you please suggest a metric to obtain the semantic similarity between sentences.

How about the Infersent and word mover's distance? I think you should consider adding these metrics for evaluation of text generation. This repository is helpful for evaluation of generated text.

@Radhikadua123
Copy link
Author

Can you please suggest which one of these metrics is widely used to get the semantic similarity between sentences?
Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants