How to limit memory usage for large datasets? #2639
Unanswered
gunnhildsp
asked this question in
Q&A
Replies: 1 comment
-
From the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Zarr version
2.16.1
Numcodecs version
0.12.1
Python Version
3.12.1
Operating System
Linux
Installation
using poetry in a virtual environment
Description
I am trying to write a zarr dataset using netcdf files. To try to limit memory usage, I am first creating daily zarr directories from hourly netcdf files, using xarray. Then I am combining the daily files into a monthly zarr directory. I finally want to write the monthly zarr to an azure blob storage. However, the process is killed (no stacktrace), I assume from running out of memory, when combining the daily zarrs to monthly. If I create a smaller final directory, for example combining two daily zarrs to one, it works fine. I am using xarray version 2024.2.0
Steps to reproduce
Additional output
No response
Beta Was this translation helpful? Give feedback.
All reactions