Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Location softmax #12

Open
pmorerio opened this issue Oct 4, 2016 · 4 comments
Open

Location softmax #12

pmorerio opened this issue Oct 4, 2016 · 4 comments

Comments

@pmorerio
Copy link

pmorerio commented Oct 4, 2016

Hi, thanks for sharing the code!
It looks like the location softmax implemented in the conditional lstm is not the one you describe in the paper 'Action Recognition with Visual Attention' (eq. 4), but rather the one described in eqns. 4,5,6,9 in 'Show, Attend and Tell: Neural Image Caption Generation with Visual Attention'. Could you please comment on that? Really appreciate, thanks!

@GerardoHH
Copy link

Hi @kracwarlock ,

I have a question, in your master thesis, in the equation 2.4, you mention that "Wi are the weights mapping to the ith element of the location softmax", I dont fully undertand this, can you explain it a little more ? thank you !

@kracwarlock
Copy link
Owner

@pmorerio the location softmax in my case is presented in (4) and in their case it is (4)+(5)
(6) and (9) hold true for both papers
the results in our paper are based on the equation in our paper's (4)
however the show attend tell paper's (4)+(5) seem to produce better results since they look at the current frame's features as well

@kracwarlock
Copy link
Owner

@GerardoHH i just meant that W_i are the weights for l_{t,i} (ith element of the location softmax).

@pmorerio
Copy link
Author

pmorerio commented Dec 19, 2016

@kracwarlock Thanks a lot for your answer. I must say that this does not look the case since in the code the location softmax is calculated base on both current features AND previous hidden state. The two contributions are then added together here and used to activate a softmax layer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants