Skip to content

Intitial condition data from open data buckets and using Kerchunk/Dynamic zarr method #11

@nishadhka

Description

@nishadhka

Thank you so much for the library and efforts put forward for simplifying the AI models on weather forecast.
Related to the initial condition data , such as in the case of NOAA GFS. For a single run (out of 00,06,12, 18), how many hours of forecasts data set (for example GFS gives in each run 240 hours of forecast for every 3 hours) is required to apply for the AI models.

Also couldn't find in the https://github.com/secondlaw-ai/skyrim/blob/master/skyrim/core/fetch.py, routine specific to GFS download as shown in documentation.

In case the intial condition dataset download is a bottleneck, is there plan to use the kerchunk https://fsspec.github.io/kerchunk/ or recent changes in using the grib index based method to stream the initial condition dataset from GFS and ensemble forecast system such as fsspec/kerchunk#399, and an GFS data stream method in https://github.com/asascience-open/nextgen-dmac/blob/main/grib_index_aggregation/dynamicgribchunking.ipynb
for GFS https://registry.opendata.aws/noaa-gfs-bdp-pds/
The initial condition dataset on ensemble forecast systems
for GEFS, https://registry.opendata.aws/noaa-gefs/
for ECMWF, https://registry.opendata.aws/ecmwf-forecasts/

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions