-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running instructions to Recreate WNED-CWEB and ACQUAINT results #2
Comments
DATAIn the data folder, we provide the datasets in "benchmark", you need to change the corresponding args to run. The data processing would take a while as the document is huge. [process via load_entity_description function in data_loader.py ] MODELFollowing BLINK, for Wikipedia-based datasets, we pre-train the BERT-large on 9M paired Wikipedia samples. Hence, the results for baseline and GER are high. To get the pre-trained BERT-large, please download the ckpt via download_blink_models.sh in https://github.com/facebookresearch/BLINK and convert it into GER format. |
What are the recommended split sizes for the two datasets in data_loader.py? Currently the WORLDS dictionary only has information about zeshel dataset
I understand that in the paper the pretrained model used was the BERT-large from BLINK, but can I run the experiments with bert-base-uncased as well without any additional changes? |
I understand that in the paper the pretrained model used was the BERT-large from BLINK, but can I run the experiments with bert-base-uncased as well without any additional changes?you can change the backbone to bert-base-uncased by changing the args and no need to change the code. |
I am trying to recreate the results from the paper. I was able to get all the ZESHEL experiment results but I was not able to find any code to get the WNED-CWEB and ACQUAINT results. Could you please give the steps to be followed to recreate these results?
The text was updated successfully, but these errors were encountered: