Cookbooks
This page contains a collection of WarpStream-specific Bento recipes to help you accomplish common tasks.
Stream to Parquet Files
input:
kafka_franz_warpstream:
topics:
- logs
output:
aws_s3:
batching:
byte_size: 32000000
count: 0
period: 5s
processors:
- mutation: |
root.value = content().string()
root.key = @kafka_key
root.kafka_topic = @kafka_topic
root.kafka_partition = @kafka_partition
root.kafka_offset = @kafka_offset
- parquet_encode:
default_compression: zstd
default_encoding: PLAIN
schema:
- name: kafka_topic
type: BYTE_ARRAY
- name: kafka_partition
type: INT64
- name: kafka_offset
type: INT64
- name: key
type: BYTE_ARRAY
- name: value
type: BYTE_ARRAY
bucket: $YOUR_S3_BUCKET
path: parquet_logs/${! timestamp_unix() }-${! uuid_v4() }.parquet
region: $YOUR_S3_REGION
warpstream:
cluster_concurrency_target: 6Stream to GCP BigQuery
Stream to GCP BigQuery Iceberg tables
Stream to Redshift
Last updated
Was this helpful?