Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training data for original model #26

Open
hasan-sayeed opened this issue Aug 19, 2021 · 4 comments
Open

Training data for original model #26

hasan-sayeed opened this issue Aug 19, 2021 · 4 comments

Comments

@hasan-sayeed
Copy link

I was wondering, is the training data used for the original model is available anywhere as the Matscholar API is currently not available.

@jdagdelen
Copy link
Contributor

jdagdelen commented Aug 19, 2021 via email

@hasan-sayeed
Copy link
Author

Could you please share the code you used to query the APIs and filter the abstracts, as described in the Methods sections of the paper?

@jdagdelen
Copy link
Contributor

jdagdelen commented Aug 20, 2021

This is still an active area of development for us, so we aren't able to release that yet. However, our code is very specific to working on our infrastructure at LBL, (database config, etc) and wouldn't really be that useful to you, I think. However, I'm happy to give you an overview of how we went about it and point you to some libraries/resources that can help.

Our code uses the pybliometrics library to connect to the ScienceDirectAPI. We constructed a list of journals we were interested in and the years they were in service from the spreadsheet published by Elsevier every year (here is the list for journals on ScienceDirect). We then split up this list into journal-year pairs and queried the ScienceDirect API for those parameters. After that, we process the entries to make sure everything has the same metadata (doe, authors, etc.) We hand-labeled a number of abstracts for relevance (I think it was 1000) and then trained a classifier, which I believe used a bag of words featurization, for relevance.

@michaeljtobias
Copy link

Regarding the hand-labels, I see there is a dois.txt file and also a relevant_dois.json, so is it correct to assume the dois.txt is the complete set, and the relevant_dois.json are those predicted in by the classifier trained with the 1000 hand-labels? In that case is it possible to provide a table of those 1000 hand labels so I could attempt to recreate the same classifier?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants