Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Removed doc, docs folder, cleanup docker compose and update titles of… #389

Merged
merged 1 commit into from
Nov 29, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -326,7 +326,7 @@ mvn install -DskipTests=true
- [Architecture](doc/architecture.md)
- [Local Setup - Docker Compose](doc/setup.md)
- [Debezium Setup](doc/debezium_setup.md)
- [Kubernetes Setup](doc/k8s_pipeline_setup.md)
- [Kubernetes Setup](doc/k8s/k8s_pipeline_setup.md)
- [Sink Configuration](doc/sink_configuration.md)
- [Testing](doc/TESTING.md)
- [Performance Benchmarking](doc/Performance.md)
Expand Down
89 changes: 0 additions & 89 deletions doc/Kafka.md

This file was deleted.

4 changes: 4 additions & 0 deletions doc/Kafka_metadata.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,7 @@
## Kafka MetaData

These are the columns that are added when the `store.kafka.metadata` configuration is enabled.

| Column | Data Type | Description |
|------------------|--------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| `_offset` | Nullable(UInt64), | Kafka offset |
Expand Down
20 changes: 0 additions & 20 deletions doc/Milestones.md

This file was deleted.

2 changes: 2 additions & 0 deletions doc/Monitoring.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,5 @@
## Sink Connector (Kafka) monitoring

Sink Connector Config
OpenJDK 11.0.14.1

Expand Down
4 changes: 3 additions & 1 deletion doc/Performance.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,6 @@
# Setup
# Load Testing using SysBench.

## Setup
We use `SysBench` to perform load testing.
https://github.com/akopytov/sysbench

Expand Down
2 changes: 1 addition & 1 deletion doc/TESTING.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
# Unit tests
The unit tests are added in the following directory. This is a work in progress.
The unit tests are added in the following directory.
[test](../src/test/com/altinity/clickhouse/sink/connector)

# Integration tests.
Expand Down
1 change: 1 addition & 0 deletions doc/Troubleshooting.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
# Troubleshooting.

### Caused by: io.debezium.DebeziumException: java.sql.SQLSyntaxErrorException: Access denied; you need (at least one of)
### the SUPER, REPLICATION CLIENT privilege(s) for this operation
Expand Down
23 changes: 1 addition & 22 deletions doc/architecture.md
Original file line number Diff line number Diff line change
@@ -1,25 +1,4 @@
# Architecture

![arch](img/arch.png)

## DeDuplicator

**Assumption**: each table has to have a **Deduplication key** specified.
Deduplication key is a set of fields, explicitly specified to be used as a basis for deduplication.
In case PK is defined it is natural to use PK as Deduplication Key as well, It is not mandatory, however, and for tables
without PK, Deduplication key is still required.

Connector maintains a map (Deduplication Map), where the key is a Deduplication key and the value is the record itself.
As soon as a new record with the same Deduplication key arrives, it either replaces the existing record
in the Deduplication Map or is dropped, based on the **Deduplication Policy**. Records from the Deduplication Map are
formed into a Batch and are flushed into the ClickHouse on either time or size-based **Dump Policy**.

> **NB** it should be noted, that time-based batch flush can not form the same batches upon replay.

Flushed rows are removed from the Deduplication Map based on either time or size-based **Clear Policy**

As a result of the Deduplication Map application, connector has a set of records, which are de-duplicated within
a certain time or size - limited window of records.
# Kafka Sink Connector Architecture

## Exactly Once Semantics

Expand Down
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
2 changes: 1 addition & 1 deletion doc/mutable_data.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
## ReplacingMergeTree
## How Updates/Deletes are Handled - ReplacingMergeTree
Sink Connector will attempt to read the `engine_full` column from system.tables for the corresponding table and will
identify the `engine` and the `ver` column.

Expand Down
2 changes: 2 additions & 0 deletions doc/postgres_wal.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,5 @@
## PostgreSQL WAL Dump utility.

Using the `pg_waldump` utility to dump the WAL log information. `pg_waldump` utility needs to be provided the postgresql data directory path.
```
pg_waldump pg_wal/000000010000000000000001
Expand Down
3 changes: 2 additions & 1 deletion doc/schema_registry.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
#### Confluent Schema Registry
# Confluent Schema Registry
### Schema Registry API calls.

REST API to get subjects
```
Expand Down
2 changes: 1 addition & 1 deletion doc/setup.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
This doc describes how to set up CDC pipeline

# Pipeline
# Setup Kafka Sink Connector Pipeline

![pipeline](img/pipeline.png)

Expand Down
3 changes: 3 additions & 0 deletions doc/sink_configuration.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,6 @@
# Kafka Sink connector configuration.


All the Sink Connector configuration variables are documented here.

`topics`: Kafka topics delimited by comma.\
Expand Down
1 change: 1 addition & 0 deletions doc/transactions.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
# MySQL Transactions(currently not supported)
Transactions have to be enabled in the source connector, For MySQL
the following flag has to be set
` "provide.transaction.metadata": "true"`
Expand Down
Loading
Loading