Skip to main content

strm create batch-exporter

Create a Batch Exporter


A Batch Exporter listens to a stream and writes all events to files using a Data Connector. This happens with a regular interval.

When exporting events each file follows the JSON Lines format, which is one full JSON document per line. When exporting encryption keys, each file is a CSV file.

A Data Connector is a configuration entity that comprises a location (GCS bucket, AWS S3 bucket, ...) and associated credentials.

A Data Connector must be created before you can create a batch exporter that uses it.


strm create batch-exporter (stream-name) [flags]


      --data-connector string     name of the data connector - optional if you own only one data connector
--export-keys do we want to export the keys stream
-h, --help help for batch-exporter
--include-existing-events do we want to include all existing events
--interval int interval in seconds between batches (default 60)
--name string optional batch exporter name (default <stream>-<dataconnector>).
--path-prefix string path prefix on bucket

Options inherited from parent commands

      --api-auth-url string   user authentication host (default "")
--api-host string api host and port (default "")
-o, --output string output format [json, json-raw, table, plain] (default "table")
-p, --project string project to use (defaults to context-configured project)
--token-file string token file that contains an access token (default is $HOME/.config/strmprivacy/credentials-<api-auth-url>.json)