-
Notifications
You must be signed in to change notification settings - Fork 101
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
entity-aware self-attention used in mLUKE #174
Comments
In our pretraining and fine-tuning experiments, the mLUKE models did not use entity-aware self-attention. The attention weights related to entity-aware self-attention (EASA) (e.g., By default, The warning is somewhat disturbing... but it is expected behavior. |
I appreciate the clarification. Now I understand.
To fix this confusing behavior, how about equipping a model with EASA weights regardless of the setting of |
Thank you for your suggestion. Other solutions I can come up with are...
As this is not the first time this kind of confusion has occurred, I will definitely do the second one later soon. |
I think we can use |
Thank you for the pointer! It is a clean solution. |
According to the mLUKE paper, mLUKE has not used entity-aware self-attention.
However, the following code gives the warning message:
(The message can be suppressed by giving "use_entity_aware_attention=True".)
In fact, the public model contains weights for entity-aware self-attention.
Could you make it clear whether mLUKE uses entity-aware self-attention?
config.json
should specifyuse_entity_aware_attention: true
unlesspytorch_model.bin
is updated with no weights for entity-aware self-attention.The text was updated successfully, but these errors were encountered: