How to limit memory usage for large datasets? #2639
Unanswered
gunnhildsp
asked this question in
Q&A
Replies: 1 comment
-
From the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Zarr version
2.16.1
Numcodecs version
0.12.1
Python Version
3.12.1
Operating System
Linux
Installation
using poetry in a virtual environment
Description
I am trying to write a zarr dataset using netcdf files. To try to limit memory usage, I am first creating daily zarr directories from hourly netcdf files, using xarray. Then I am combining the daily files into a monthly zarr directory. I finally want to write the monthly zarr to an azure blob storage. However, the process is killed (no stacktrace), I assume from running out of memory, when combining the daily zarrs to monthly. If I create a smaller final directory, for example combining two daily zarrs to one, it works fine. I am using xarray version 2024.2.0
Steps to reproduce
Additional output
No response
Beta Was this translation helpful? Give feedback.
All reactions