Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Duckdb mysql extension took extreme longer time to access data than other components #95

Open
dawsondong opened this issue Oct 31, 2024 · 0 comments

Comments

@dawsondong
Copy link

Based on such TPCH test background:
Table: orders
Data Amount: 150,000,000
SQL to use duckdb mysql extension: copy my.orders to '/tmp/orders.mysql.duckdb.parquet' (format parquet);
Time to consume: 534 seconds.

SQL to use Doris: select * from my.test.orders into outfile "file:///tmp/orders.mysql.parquet" format as parquet;
Time to consume: 122 seconds.

With simple JDBC application to access orders table and write, it will consume about 150 seconds which is closed to Doris.

So in most of cases, duckdb_mysql takes over 4 times cost to achieve same task. It could get improved.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant