Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

如果有limit的情况下,根据source row count来判断是否一次取数 #96

Open
hn5092 opened this issue Jan 7, 2020 · 0 comments
Assignees

Comments

@hn5092
Copy link

hn5092 commented Jan 7, 2020

https://github.com/Kyligence/KAP/issues/17384 因为bitmap很大但是行数很少,如果每次还是一个分区一个分区取数很导致查询变慢

@hn5092 hn5092 self-assigned this Jan 7, 2020
chenzhx pushed a commit to chenzhx/spark that referenced this issue Feb 24, 2022
…in parquet

### What changes were proposed in this pull request?
Spark should remove check field name when reading/writing  parquet files.

### Why are the changes needed?
Support spark reading existing parquet files with special chars in column names.

### Does this PR introduce _any_ user-facing change?
Such as  parquet, user can use spark to read existing files with special chars in column names. And then can use back quote  to wrap special column name such as  &Kyligence#96;max(t)&Kyligence#96; or use &Kyligence#96;max(t)&Kyligence#96; as &Kyligence#96;max_t&Kyligence#96;, then user can use `max_t`.

### How was this patch tested?

Added UT

Closes apache#35229 from AngersZhuuuu/SPARK-27442.

Authored-by: Angerszhuuuu <angers.zhu@gmail.com>
Signed-off-by: Wenchen Fan <wenchen@databricks.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant