-
Notifications
You must be signed in to change notification settings - Fork 47
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
which attention architecture is used in NER? #6
Comments
According to the README, attention mechanism not suitable for NER task: |
In my own experiment, attention mechanism really not work, but Layer Normalization can improve the robustness of the model |
well in my experiments attention only models achieve comparable results to LSTM, even got better than LSTM with very less training time |
Maybe there are some problems with my experiment. I have tried BLSTM+SelfAttention+CRF,the effect is not as good as BLSTM+CRF. The structure of your model is SelfAttention+CRF, not LSTM? I want to give it a try. |
yes, structure of my model is attention+crf only |
do you have the relevant code, can I refer to it? |
@MingLunHan excuse me, what is your model's architecture? bilstm + attention only? |
I want to understand how you used attention in NEr task, any paper or article which explains this? Thanks
The text was updated successfully, but these errors were encountered: