:::danger
This destination is meant to be used on a local workstation and won't work on Kubernetes
:::
This destination writes data to a directory on the local filesystem on the host running Airbyte. By default, data is written to /tmp/airbyte_local
. To change this location, modify the LOCAL_ROOT
environment variable for Airbyte.
Each stream will be output into its own file. Each file will contain 3 columns:
_airbyte_ab_id
: a uuid assigned by Airbyte to each event that is processed._airbyte_emitted_at
: a timestamp representing when the event was pulled from the data source._airbyte_data
: a json blob representing with the event data.
Feature | Supported | |
---|---|---|
Full Refresh Sync | Yes | |
Incremental - Append Sync | Yes | |
Incremental - Deduped History | No | As this connector does not support dbt, we don't support this sync mode on this destination. |
Namespaces | No |
This integration will be constrained by the speed at which your filesystem accepts writes.
The destination_path
will always start with /local
whether it is specified by the user or not. Any directory nesting within local will be mapped onto the local mount.
By default, the LOCAL_ROOT
env variable in the .env
file is set /tmp/airbyte_local
.
The local mount is mounted by Docker onto LOCAL_ROOT
. This means the /local
is substituted by /tmp/airbyte_local
by default.
- If
destination_path
is set to/local/cars/models
- the local mount is using the
/tmp/airbyte_local
default - then all data will be written to
/tmp/airbyte_local/cars/models
directory.
If your Airbyte instance is running on the same computer that you are navigating with, you can open your browser and enter file:///tmp/airbyte_local to look at the replicated data locally. If the first approach fails or if your Airbyte instance is running on a remote server, follow the following steps to access the replicated files:
- Access the scheduler container using
docker exec -it airbyte-scheduler bash
- Navigate to the default local mount using
cd /tmp/airbyte_local
- Navigate to the replicated file directory you specified when you created the destination, using
cd /{destination_path}
- List files containing the replicated data using
ls
- Execute
cat {filename}
to display the data in a particular file
You can also copy the output file to your host machine, the following command will copy the file to the current working directory you are using:
docker cp airbyte-scheduler:/tmp/airbyte_local/{destination_path}/{filename}.csv .
Note: If you are running Airbyte on Windows with Docker backed by WSL2, you have to use similar step as above or refer to this link for an alternative approach.