Skip to content
This repository has been archived by the owner on Apr 11, 2021. It is now read-only.

The incorporation of attention in attention_lstm.py #25

Open
ns-moosavi opened this issue Sep 24, 2016 · 0 comments
Open

The incorporation of attention in attention_lstm.py #25

ns-moosavi opened this issue Sep 24, 2016 · 0 comments

Comments

@ns-moosavi
Copy link

In the blog post and in the related literature about attention LSTM, attention is incorporated like

attention_state = tanh(dot(attention_vec, W_attn) + dot(new_hidden_state, U_attn))

However, in attention_lstm.py it is incorporated like:

attention_state = tanh(dot(attention_vec, W_attn) * dot(new_hidden_state, U_attn))

Is it a typo or do you find it a better way of incorporating attention?

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant