There are 0 repository under amazon-msk topic.
Sample code that shows the important aspects of developing custom connectors for Kafka Connect. It provides the resources for building, deploying, and running the code on-premises using Docker, as well as running the code in the cloud.
Managed ELKK stack implemented with the AWS CDK
Technical solution to implement event processing prioritization with Apache Kafke using the concept of buckets.
After you migrate from an existing graph database to Amazon Neptune, you might want to capture and process changed data in real time. Continuous replication of databases using the change data capture technique allows you to unlock your data and make it available to other systems for use cases such as distributed data processing, building an enterprise data lake, and modernizing your existing database. In the previous post of this series, we demonstrated with an example solution, how to perform automated migration from Neo4j to Amazon Neptune. If you are looking beyond one-time migration and want to keep both databases in sync you might want to run an ongoing replication using the change data capture technique.
Kafka KStreams example using Amazon MSK with IAM Auth
Data Pipeline for CDC data from MySQL DB to Amazon S3 through Amazon MSK using Amazon MSK Connect (Debezium).
AWS Secrets Manager Config Provider for Apache Kafka
A tool that can be deployed to process posting and receiving text and audio files from and into a data lake, apply transformation in a distributed manner, and load it into a warehouse in a suitable format to train a speech-to-text model
Demo event analytics platform based on Apache Kafka (Confluent).