Spring Cloud Data Flow (SCDF) is a data integration and orchestration service for composing, deploying, and managing data pipelines.
- Streaming pipelines for event-driven use cases (real-time ETL, messaging, analytics).
- Task pipelines for batch or scheduled workloads (machine learning jobs, database migrations, reporting).
- Application orchestration across multiple runtimes: Cloud Foundry, Kubernetes, or Local.
- Scalability with partitioning, scaling, and monitoring support.
A pipeline in SCDF is built from Spring Cloud Stream and Spring Cloud Task applications, typically composed of:
- Source – Ingests data (e.g., from Kafka, RabbitMQ, HTTP, File, JDBC).
- Processor – Transforms or enriches data.
- Sink – Writes data to a target system (e.g., database, messaging system, file, analytics store).
| Application | Notes |
|---|---|
| account-batch | Spring Batch/Task Example |
| postgres-query-processor | Postgres Based SQL Streaming Processor |
| postgres-sink | Postgres Based SQL Streaming Sink |
See Hands On Labs