Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I couldn't find any measure of query difficulty in the Spider dataset #83

Open
Atlamtiz opened this issue Feb 24, 2023 · 3 comments
Open

Comments

@Atlamtiz
Copy link

"I didn't find any difficulty metric in Spider, but in the latest paper, Graphix-T5: Mixing Pre-Trained Transformers with Graph-Aware Layers for Text-to-SQL Parsing, I saw that they distinguished different difficulty levels.
However, there seems to be no difficulty measurement in the dataset. Why is that?"

@hanlinGao
Copy link

hanlinGao commented Mar 3, 2023

I have the same confusion. I classified the samples in dev.json based on the criteria defined in README, but when I input my results in the evaluation.py, I found the evaluation.py will classify the gold.txt into different hardness and the classification results were slightly different from mine.

So I went to the evaluation.py and used the counting functions there instead to classify the samples

@wanjianwei
Copy link

+1

@iamlockelightning
Copy link

👀

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants