-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Task import-heat-demand fails with --dataset-boundary=Everything #204
Comments
I do not have a deep understanding of the reported problem. I just want to tell you, that I successfully run the heat demand data import for Germany, before we implemented the test case and the new docker container (and a few heat demand table related changes e.g. version numbering) on my laptop. Maybe this hint helps. |
Thanks for the hint, @EvaWie . This is strange. It seems that the culprit is the line importing the rasters. This executes a SQL script which inserts a row that is simply to large for PostgreSQL. The whole raster_data is saved in one row and single objects like row entries are only allowed to be up to 1GB (=1073741824 Bytes) big, which the value in the log messages exceeds. So we either have to find a way of splitting the rasters into more than one row, or we'll have to find a way of storing the rasters as "large objects". When implementing the raster import, I actually experimented a bit with storing the rasters in more than one row, so I might be able to dig up some code which could help here. |
I remember that we have changed the datatype from int to float. If that was done after Evas run, this could cause the difference and also the new issue. I will check if it works to import the values as integers. |
That is a very good point! |
I tried to run it with Integer values and it worked. So we know at least why this error didn't occurred before. |
Nice.
It's not much. I think it would be enough to use the |
…erything Fixes/#204 heat demand import everything
The heat demand import fails with the following log, the log file is ~4GB so I wasn't even able to open it completely, but this screenshot shows the most important lines (just the huge list of values is missing).
It looks the these lines need too much memory:
eGon-data/src/egon/data/importing/heat_demand_data/__init__.py
Lines 478 to 489 in 64b7921
I think we can't really change something in these lines to reduce the memory because it is already directly inserted to the database.
But regarding htop, the maximal RAM used is ~20GB and the server has much more.
@gnn Is there a limit for airflow or the docker container and could we try to adjust this?
The text was updated successfully, but these errors were encountered: