Kafka topic replication factor
Kafka replication helps prevent data loss by writing the same data to more than one broker.
- A replication factor of
1
means no replication. It is mostly used for development purposes and should be avoided in test and production Kafka clusters - A replication factor of
3
is a commonly used replication factor as it provides the right balance between broker loss and replication overhead.
2
. When a message is written down into Partition 0 of Topic-A in Broker 101, it is also written down into Broker 102 because it has Partition 0 as a replica.


What are Kafka partitions leader and replicas?
For a given topic-partition, one Kafka broker is designated by the cluster to be responsible for sending and receiving data to clients. That broker is known as the leader broker of that topic partition. Any other broker that is storing replicated data for that partition is referred to as a replica. Therefore, each partition has one leader and multiple replicas.What are in-sync replicas (ISR)?
An ISR is a replica that is up to date with the leader broker for a partition. Any replica that is not up to date with the leader is out of sync.
Kafka producers acks setting
Kafka producers only write data to the current leader broker for a partition. Kafka producers must also specify a level of acknowledgmentacks
to specify if the message must be written to a minimum number of replicas before being considered a successful write.
The default value of
acks
has changed in Kafka v3.0:- if using Kafka < v3.0,
acks=1
- if using Kafka >= v3.0,
acks=all
acks=0
Whenacks=0
producers consider messages as “written successfully” the moment the message was sent without waiting for the broker to accept it at all.

acks = 1
Whenacks=1
, producers consider messages as “written successfully” when the message was acknowledged by only the leader.

acks = all
Whenacks=all
, producers consider messages as “written successfully” when the message is accepted by all in-sync replicas (ISR).

min.insync.replicas
). The request will be stored in a buffer until the leader observes that the follower replicas replicated the message, at which point a successful acknowledgement is sent back to the client.
Themin.insync.replicas
can be configured both at the topic and the broker-level. The data is considered committed when it is written to all in-sync replicas - min.insync.replicas.
A value of 2 implies that at least 2 brokers that are ISR (including leader) must respond that they have the data.
If you would like to be sure that committed data is written to more than one replica, you need to set the minimum number of in-sync replicas to a higher value. If a topic has three replicas and you set min.insync.replicas
to 2
, then you can only write to a partition in the topic if at least two out of the three replicas are in-sync. When all three replicas are in-sync, everything proceeds normally. This is also true if one of the replicas becomes unavailable. However, if two out of three replicas are not available, the brokers will no longer accept produce requests. Instead, producers that attempt to send data will receive NotEnoughReplicasException
.

Kafka topic durability and availability
For a topic replication factor of 3, topic data durability can withstand the loss of 2 brokers. As a general rule, for a replication factor ofN
, you can permanently lose up to N-1
brokers and still recover your data.
Regarding availability, it is a little bit more complicated… To illustrate, let’s consider a replication factor of 3:
- Reads: As long as one partition is up and considered an ISR, the topic will be available for reads
- Writers:
acks=0
&acks=1
: as long as one partition is up and considered an ISR, the topic will be available for writes.acks=all
:min.insync.replicas=1
(default): the topic must have at least 1 partition up as an ISR (that includes the reader) and so we can tolerate two brokers being downmin.insync.replicas=2
: the topic must have at least 2 ISR up, and therefore we can tolerate at most one broker being down (in the case of replication factor of 3), and we have the guarantee that for every write, the data will be at least written twice.min.insync.replicas=3
: this wouldn’t make much sense for a corresponding replication factor of 3 and we couldn’t tolerate any broker going down.- in summary, when
acks=all
with areplication.factor=N
andmin.insync.replicas=M
we can tolerateN-M
brokers going down for topic availability purposes
Kafka topic replication settings
acks=all
and min.insync.replicas=2
is the most popular option for data durability and availability and allows you to withstand at most the loss of one Kafka broker.Kafka consumers replicas fetching
Kafka consumers read by default from the partition leader.

Preferred leader
The preferred leader is the designated leader broker for a partition at topic creation time (as opposed to being a replica).Leader election is the process of deciding which broker is a leader at topic creation time is called a preferred leader election.