Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Motivation for x_context #1

Open
neverfox opened this issue May 20, 2018 · 4 comments
Open

Motivation for x_context #1

neverfox opened this issue May 20, 2018 · 4 comments
Labels
question Further information is requested

Comments

@neverfox
Copy link

I am curious about the motivation for this step, as I couldn't anything about it in See et al., though I might have just missed it.

@atulkum
Copy link
Owner

atulkum commented May 20, 2018

Yes in the paper it is not mentioned anywhere but the code has it.

https://github.com/abisee/pointer-generator/blob/master/attention_decoder.py#L150

@atulkum atulkum added the question Further information is requested label May 20, 2018
@atulkum
Copy link
Owner

atulkum commented Jul 20, 2018

I get the paper where similar kind of attention mechanism is used.

Order Matters: Sequence to sequence for sets

@lorenzoale
Copy link

Thanks a lot for this work!
I have a question about when do you update this context. During training, as far as I understand, seems to be ok. However, during decoding, in the abisee code they seem to update the context before using it in lstm:
https://github.com/abisee/pointer-generator/blob/a7317f573d01b944c31a76bde7218bcfc890ef6a/attention_decoder.py#L140
And then also update it after the call to lstm.
https://github.com/abisee/pointer-generator/blob/a7317f573d01b944c31a76bde7218bcfc890ef6a/attention_decoder.py#L158

@atulkum
Copy link
Owner

atulkum commented Oct 2, 2018

Thanks for pointing this out. You are right. I have updated my code, I still need to re-run the experiments though. I will update the result after that. Here is the commit:
454a2f6
By the way I am thinking about experimenting with transformer network for this task and verify speed and accuracy tradeoff. Let me know if you want to collaborate.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants