-
Notifications
You must be signed in to change notification settings - Fork 376
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Attention type #165
Comments
actually after double checking it, it looks like its the dot attention of Luong. Is there a reason that use the dot attention and not the general one? |
@ratis86 thanks for pointing this out. There's no particular reason that I'm aware of. You can contact the respective contributor for that. However we're gonna be implementing the general as well as copy attention mechanisms in the coming versions. |
@pskrunner14 And also on this. Whom should I contact? |
@CoderINusE you're welcome to submit a PR. |
@pskrunner14 should I pass an additional argument to the |
@CoderINusE please see |
I am not sure whether the comment section in current Attention Module is a bit off? "output=tanh(w∗(attn∗context)+b∗output)" does not match with the code or the 5th equation in the paper https://arxiv.org/pdf/1508.04025.pdf unless b is also interpreted as a matrix? Thanks |
I think there is a difference between math written in comments and code. |
Can somebody tell me what is the type of attention used in this lib? Because I checked against Bahdanau and Luong attentions and it doesn't look like neither or maybe I'm missing something !
The text was updated successfully, but these errors were encountered: