Storing large data frames #575
Replies: 4 comments
-
Thank you for this interesting feedback. For a warning, I am not sure what the right behavior would be. So I suggest picking sane defaults:
Then an enhancement that must examined in a wider context:
|
Beta Was this translation helpful? Give feedback.
-
As suggested by @rouk1, we could think about using Perspective as a visualization tool unlocking many interesting use-cases, including the ability to store large amounts of data in a skore database.. |
Beta Was this translation helpful? Give feedback.
-
Discussed with @fcharras that suggested to have a look at DeltaLake, and parquet files. |
Beta Was this translation helpful? Give feedback.
-
It would be nice to allow storing large dataframes so that users can share data and benefit from skore's ability to store successive versions. Data over a certain size could be stored as parquet files. |
Beta Was this translation helpful? Give feedback.
-
Issue
From the notebook, suppose that I load a large dataframe of shape (1M, 1k) let's say. I try to store it: it takes a lot of time.
Expectation
I would like a warning to the user that his file is too big, or maybe just store a subset of the big Dataframe (and warning that the display of the dashboard is truncated). Indeed, no one needs to visualize a dataframe of shape (1M, 1k) in the dashboard.
Beta Was this translation helpful? Give feedback.
All reactions