-
Notifications
You must be signed in to change notification settings - Fork 11
Description
Thank you so much for the library and efforts put forward for simplifying the AI models on weather forecast.
Related to the initial condition data , such as in the case of NOAA GFS. For a single run (out of 00,06,12, 18), how many hours of forecasts data set (for example GFS gives in each run 240 hours of forecast for every 3 hours) is required to apply for the AI models.
Also couldn't find in the https://github.com/secondlaw-ai/skyrim/blob/master/skyrim/core/fetch.py, routine specific to GFS download as shown in documentation.
In case the intial condition dataset download is a bottleneck, is there plan to use the kerchunk https://fsspec.github.io/kerchunk/ or recent changes in using the grib index based method to stream the initial condition dataset from GFS and ensemble forecast system such as fsspec/kerchunk#399, and an GFS data stream method in https://github.com/asascience-open/nextgen-dmac/blob/main/grib_index_aggregation/dynamicgribchunking.ipynb
for GFS https://registry.opendata.aws/noaa-gfs-bdp-pds/
The initial condition dataset on ensemble forecast systems
for GEFS, https://registry.opendata.aws/noaa-gefs/
for ECMWF, https://registry.opendata.aws/ecmwf-forecasts/