This project is an ETL pipeline designed to extract weather data from the Open-Meteo API, transform it using Python and Pandas, and load it into storage using Apache Airflow for orchestration.
We are fetching real-time weather data using the Open-Meteo API.
https://api.open-meteo.com/v1/forecast?latitude=51.5074&longitude=-0.1278¤t_weather=true
- Data Source: Open-Meteo API for weather data.
- ETL Script: Python script to extract, transform (clean using Pandas), and load data.
- Orchestration: Apache Airflow to schedule and monitor ETL workflows.
- Storage: Postgres database and optional CSV backups.
- Cloud Integration: AWS (for extended storage or deployment).
├── dags/
│ └── ETL_DataPipeline.py
├── Dockerfile
├── docker-compose.yaml
├── README.md
└── requirements.txt
- Clone the repository.
- Build and start Airflow:
docker compose up -d --build
- Access Airflow UI at http://localhost:8085.
- Unpause the
weather_etl_pipeline
DAG and trigger it.
- Raghav Tigadi