Singer tap for .DBF files
Project description
tap-dbf
Singer tap for the dBase file format.
Configuration
Setting | Required | Default | Description |
---|---|---|---|
path | True | None | Glob expression where the files are located. Stream names will be extracted from the file name. |
fs_root | False | file:// | The root of the filesystem to read from. |
ignore_missing_memofile | False | 0 | Whether to proceed reading the file even if the memofile is not present. |
s3 | False | None | S3 configuration. |
s3.key | False | None | The AWS key ID. |
s3.secret | False | None | The AWS secret key. |
s3.endpoint_url | False | None | The S3 endpoint URL. |
gcs | False | None | GCS configuration. |
gcs.token | False | None | OAuth 2.0 token for GCS. |
stream_maps | False | None | Config object for stream maps capability. For more information check out Stream Maps. |
stream_map_config | False | None | User-defined config values to be used within map expressions. |
faker_config | False | None | Config for the Faker instance variable fake used within map expressions. Only applicable if the plugin specifies faker as an addtional dependency (through the singer-sdk faker extra or directly). |
faker_config.seed | False | None | Value to seed the Faker generator for deterministic output: https://faker.readthedocs.io/en/master/#seeding-the-generator |
faker_config.locale | False | None | One or more LCID locale strings to produce localized output for: https://faker.readthedocs.io/en/master/#localization |
flattening_enabled | False | None | 'True' to enable schema flattening and automatically expand nested properties. |
flattening_max_depth | False | None | The max depth to flatten schemas. |
batch_config | False | None | |
batch_config.encoding | False | None | Specifies the format and compression of the batch files. |
batch_config.encoding.format | False | None | Format to use for batch files. |
batch_config.encoding.compression | False | None | Compression format to use for batch files. |
batch_config.storage | False | None | Defines the storage layer to use when writing batch files |
batch_config.storage.root | False | None | Root path to use when writing batch files. |
batch_config.storage.prefix | False | None | Prefix to use when writing batch files. |
JSON example
{
"path": "tests/data/files/*.dbf",
"ignore_missing_memofile": true
}
Filesystems
Local
Example configuration:
{
"path": "/files/*.dbf",
"fs_root": "file://data",
"ignore_missing_memofile": true
}
The fs_root
key is optional and defaults to the current working directory:
{
"path": "data/files/*.dbf",
"ignore_missing_memofile": true
}
S3
You need to install the package with the s3
extra:
pip install 'tap-dbf[s3]'
Example configuration:
{
"path": "/*.dbf",
"fs_root": "s3://files",
"ignore_missing_memofile": true,
"s3": {
"key": "someKey",
"secret": "someSecret",
"endpoint_url": "http://localhost:9000"
}
}
Google Cloud Storage
You need to install the package with the gcs
extra:
pip install 'tap-dbf[gcs]'
Example configuration:
{
"path": "/*.dbf",
"fs_root": "gcs://files",
"ignore_missing_memofile": true,
"gcs": {
"token": "cloud"
}
}
See https://gcsfs.readthedocs.io/en/latest/#credentials for more information about the token
key.
Roadmap
- Google Drive filesystem
- Dropbox filesystem
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
tap_dbf-0.1.4.tar.gz
(15.5 kB
view hashes)
Built Distribution
tap_dbf-0.1.4-py3-none-any.whl
(11.6 kB
view hashes)