Singer tap for the dBase file format.
Setting | Required | Default | Description |
---|---|---|---|
path | True | None | Glob expression where the files are located. Stream names will be extracted from the file name. |
fs_root | False | file:// | The root of the filesystem to read from. |
ignore_missing_memofile | False | 0 | Whether to proceed reading the file even if the memofile is not present. |
s3 | False | None | S3 configuration. |
s3.key | False | None | The AWS key ID. |
s3.secret | False | None | The AWS secret key. |
s3.endpoint_url | False | None | The S3 endpoint URL. |
gcs | False | None | GCS configuration. |
gcs.token | False | None | OAuth 2.0 token for GCS. |
stream_maps | False | None | Config object for stream maps capability. For more information check out Stream Maps. |
stream_map_config | False | None | User-defined config values to be used within map expressions. |
faker_config | False | None | Config for the Faker instance variable fake used within map expressions. Only applicable if the plugin specifies faker as an addtional dependency (through the singer-sdk faker extra or directly). |
faker_config.seed | False | None | Value to seed the Faker generator for deterministic output: https://faker.readthedocs.io/en/master/#seeding-the-generator |
faker_config.locale | False | None | One or more LCID locale strings to produce localized output for: https://faker.readthedocs.io/en/master/#localization |
flattening_enabled | False | None | 'True' to enable schema flattening and automatically expand nested properties. |
flattening_max_depth | False | None | The max depth to flatten schemas. |
batch_config | False | None | |
batch_config.encoding | False | None | Specifies the format and compression of the batch files. |
batch_config.encoding.format | False | None | Format to use for batch files. |
batch_config.encoding.compression | False | None | Compression format to use for batch files. |
batch_config.storage | False | None | Defines the storage layer to use when writing batch files |
batch_config.storage.root | False | None | Root path to use when writing batch files. |
batch_config.storage.prefix | False | None | Prefix to use when writing batch files. |
{
"path": "tests/data/files/*.dbf",
"ignore_missing_memofile": true
}
Example configuration:
{
"path": "/files/*.dbf",
"fs_root": "file://data",
"ignore_missing_memofile": true
}
The fs_root
key is optional and defaults to the current working directory:
{
"path": "data/files/*.dbf",
"ignore_missing_memofile": true
}
You need to install the package with the s3
extra:
pip install 'tap-dbf[s3]'
Example configuration:
{
"path": "/*.dbf",
"fs_root": "s3://files",
"ignore_missing_memofile": true,
"s3": {
"key": "someKey",
"secret": "someSecret",
"endpoint_url": "http://localhost:9000"
}
}
You need to install the package with the gcs
extra:
pip install 'tap-dbf[gcs]'
Example configuration:
{
"path": "/*.dbf",
"fs_root": "gcs://files",
"ignore_missing_memofile": true,
"gcs": {
"token": "cloud"
}
}
See https://gcsfs.readthedocs.io/en/latest/#credentials for more information about the token
key.
- Google Drive filesystem
- Dropbox filesystem