Effective Design Patterns for a Kafka Cluster
A guide to implement effective Kafka Clusters Design Strategies using Partitioning and Replication
At the lowest denomination level may be created with a single broker instance. Using a Kafka Producer Data Stream can be sent in form of messages to the Kafka Broker. These messages stay on the Kafka Broker for a configurable period of time until a Kafka Consumer can retrieve and process them.
Looks like a pretty simple design…but has a few drawbacks
What if the Kafka Broker is not able to keep up with the high traffic demands?
The solution is simple, break your Kafka Topic into multiple Partitions, configure multiple Brokers and save each partition on a separate Broker. Therefore instead of one broker taking the entire load you now have many of them sharing the load. The question is how many partitions are optimal? For a successful Leader election by…