Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How about results #4

Open
xyznlp opened this issue Dec 5, 2018 · 5 comments
Open

How about results #4

xyznlp opened this issue Dec 5, 2018 · 5 comments

Comments

@xyznlp
Copy link

xyznlp commented Dec 5, 2018

Good job. How about rouge scores?

@astariul
Copy link

astariul commented Mar 7, 2019

I tried to reproduce results reported in the paper Get to the Point, using same parameters.

But I was unsuccessful, results are much lower than expected, the ROUGE scores are half lower than the ones reported in the paper.

Any idea / pretrained model available ?

@xcfcode
Copy link

xcfcode commented Apr 23, 2019

@colanim Have you got good results?

@astariul
Copy link

Nope, I gave up and tried this repo, which is really easy to reproduce and gave better results anyway.

@xcfcode
Copy link

xcfcode commented Apr 23, 2019

@colanim Thanks a lot!

@xcfcode
Copy link

xcfcode commented Apr 24, 2019

@colanim @ymfa I have changed something which can help get better results.
@mollynatsu Maybe someone can try to reproduce the results, This is an excellent repo, do not give up.
After which I can get ROUGE-L F1 score about 30

I hope this can help the one who will hold this repo in the future.

  • The input of decoder should be (word embedding + t-1 time context vector )

output, hidden = self.gru(embedded.unsqueeze(0), hidden) # unsqueeze and squeeze are necessary

  • The ptr calculation should consider the word embedding

prob_ptr = F.sigmoid(self.ptr(combined)) # (batch size, 1)

  • Init GRU params
    elif isinstance(m, nn.GRU):
        print("Init GRU params...... ")
        for param in m.parameters():
            if len(param.shape) >= 2:
                init.orthogonal_(param.data)
            else:
                init.normal_(param.data)
  • I use adam with lr equal to 0.001
  • Change attention to MLP attention and create mask
  • norm the loss by traget_len

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants