Singer tap for the dBase file format.
| Setting | Required | Default | Description |
|---|---|---|---|
| path | True | None | Glob expression where the files are located. Stream names will be extracted from the file name. |
| fs_root | False | file:// | The root of the filesystem to read from. |
| ignore_missing_memofile | False | 0 | Whether to proceed reading the file even if the memofile is not present. |
| s3 | False | None | S3 configuration. |
| s3.key | False | None | The AWS key ID. |
| s3.secret | False | None | The AWS secret key. |
| s3.endpoint_url | False | None | The S3 endpoint URL. |
| gcs | False | None | GCS configuration. |
| gcs.token | False | None | OAuth 2.0 token for GCS. |
| stream_maps | False | None | Config object for stream maps capability. For more information check out Stream Maps. |
| stream_map_config | False | None | User-defined config values to be used within map expressions. |
| faker_config | False | None | Config for the Faker instance variable fake used within map expressions. Only applicable if the plugin specifies faker as an addtional dependency (through the singer-sdk faker extra or directly). |
| faker_config.seed | False | None | Value to seed the Faker generator for deterministic output: https://faker.readthedocs.io/en/master/#seeding-the-generator |
| faker_config.locale | False | None | One or more LCID locale strings to produce localized output for: https://faker.readthedocs.io/en/master/#localization |
| flattening_enabled | False | None | 'True' to enable schema flattening and automatically expand nested properties. |
| flattening_max_depth | False | None | The max depth to flatten schemas. |
| batch_config | False | None | |
| batch_config.encoding | False | None | Specifies the format and compression of the batch files. |
| batch_config.encoding.format | False | None | Format to use for batch files. |
| batch_config.encoding.compression | False | None | Compression format to use for batch files. |
| batch_config.storage | False | None | Defines the storage layer to use when writing batch files |
| batch_config.storage.root | False | None | Root path to use when writing batch files. |
| batch_config.storage.prefix | False | None | Prefix to use when writing batch files. |
{
"path": "tests/data/files/*.dbf",
"ignore_missing_memofile": true
}Example configuration:
{
"path": "/files/*.dbf",
"fs_root": "file://data",
"ignore_missing_memofile": true
}The fs_root key is optional and defaults to the current working directory:
{
"path": "data/files/*.dbf",
"ignore_missing_memofile": true
}You need to install the package with the s3 extra:
pip install 'tap-dbf[s3]'Example configuration:
{
"path": "/*.dbf",
"fs_root": "s3://files",
"ignore_missing_memofile": true,
"s3": {
"key": "someKey",
"secret": "someSecret",
"endpoint_url": "http://localhost:9000"
}
}You need to install the package with the gcs extra:
pip install 'tap-dbf[gcs]'Example configuration:
{
"path": "/*.dbf",
"fs_root": "gcs://files",
"ignore_missing_memofile": true,
"gcs": {
"token": "cloud"
}
}See https://gcsfs.readthedocs.io/en/latest/#credentials for more information about the token key.
- Google Drive filesystem
- Dropbox filesystem