sjwiesman / flink-parquet-writer

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Flink Parquet Writer

A simple Flink job that reads Avro records from Kafka and writes the data as Hive partitioned parquet on MinIO. This playground environment includes:

  • Flink Job
  • Kafka
  • Confluent Schema Registry
  • Zookeeper (for Kafka)
  • A data generator
  • MinIO

Usage

$ docker-compose build
$ docker-compose up -d

You can then navigate localhost:9000 to see the output data. The access key is demo-key and the access secret is demo-secret.

About


Languages

Language:Java 86.3%Language:Dockerfile 11.3%Language:Shell 2.3%