-
Notifications
You must be signed in to change notification settings - Fork 19.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multi Head Attention Layer #7803
Comments
Thank @andhus I'll follow the updates |
Yeah attention layer is defacto standard used in NLP problems to achieve state of art be it generative or classification. I have implemented the attention layer in keras, and have obtained good results from it. It could be be much better if the layer is added to keras, so public can directly use it. Should I share the implementation in the thread or how is the procedure ? |
@soham97 it would be great if you could i was trying to implement it couldn't get the thing working |
Hi, I implemented this some time ago in a fork. It is somewhat dirty and lacks test suites, but it works (an NMT example of this). Cheers. |
@lvapeab thanks, man !! |
Whats the status on this? |
Closing, as there is now a Keras-friendly multi-head attention layer in TensorFlow Addons. Thanks for the feature request! |
I think is a good idea start to think how to implement this sort of layer in Keras.
I know that is a really fresh algorithm, but I believe that's a new cutting edge tech in Deep Learning for the next years.
Paper: Attention is all you need (https://arxiv.org/abs/1706.03762)
Blog showing some results: Google Research Blog
Tensor2Tensor library tensor2tensor
Pytorch implementation pytorch-t2t
The text was updated successfully, but these errors were encountered: