A new platform for documentation and tutorials is launching soon.
We are migrating Consul documentation into HashiCorp Developer, our new developer experience.
»Manually Bootstrapping a Datacenter
When deploying Consul to a datacenter for the first time, there is an initial bootstrapping that must be done. As of Consul 0.4, an automatic bootstrapping is available and is the recommended approach. However, older versions only support a manual bootstrap that is documented here.
Generally, the first nodes that are started are the server nodes. Remember that an agent can run in both client and server mode. Server nodes are responsible for running the consensus protocol, and storing the cluster state. The client nodes are mostly stateless and rely on the server nodes, so they can be started easily.
Manual bootstrapping requires that the first server that is deployed in a new
datacenter provide the
-bootstrap configuration option.
This option allows the server
to assert leadership of the cluster without agreement from any other server.
This is necessary because at this point, there are no other servers running in
the datacenter! Lets call this first server
Node A. When starting
something like the following will be logged:
2014/02/22 19:23:32 [INFO] consul: cluster leadership acquired
Node A is running, we can start the next set of servers. There is a
deployment table that covers various
options, but it is recommended to have 3 or 5 total servers per datacenter. A
single server deployment is highly discouraged as data loss is inevitable
in a failure scenario. We start the next servers without specifying
-bootstrap. This is critical, since only one server should ever be running in
bootstrap mode. Once
Node B and
Node C are started, you should see a
message to the effect of:
[WARN] raft: EnableSingleNode disabled, and no known peers. Aborting election.
This indicates that the node is not in bootstrap mode, and it will not elect
itself as leader. We can now join these machines together. Since a join
operation is symmetric it does not matter which node initiates it. From
Node B and
Node C you can do the following:
$ consul join <Node A Address> Successfully joined cluster by contacting 1 nodes.
Node A you can do the following:
$ consul join <Node B Address> <Node C Address> Successfully joined cluster by contacting 2 nodes.
Once the join is successful,
Node A should output something like:
[INFO] raft: Added peer 127.0.0.2:8300, starting replication .... [INFO] raft: Added peer 127.0.0.3:8300, starting replication
As a sanity check, the
consul info command is a useful tool. It can be used to
raft.num_peers is now 2, and you can view the latest log index under
raft.last_log_index. When running
consul info on the followers, you should
raft.last_log_index converge to the same value as the leader begins
replication. That value represents the last log entry that has been stored on
This indicates that
Node B and
Node C have been added as peers. At this
point, all three nodes see each other as peers,
Node A is the leader, and
replication should be working.
The final step is to remove the
-bootstrap flag. This is important since we
don't want the node to be able to make unilateral decisions in the case of a
failure of the other two nodes. To do this, we send a
Node A to
allow it to perform a graceful leave. Then we remove the
-bootstrap flag and
restart the node. The node will need to rejoin the cluster, since the graceful
exit leaves the cluster. Any transactions that took place while
Node A was
offline will be replicated and the node will catch up.
Now that the servers are all started and replicating to each other, all the
remaining clients can be joined. Clients are much easier, as they can be started
and perform a
join against any existing node. All nodes participate in a
gossip protocol to perform basic discovery, so clients will automatically find
the servers and register themselves.
If you accidentally start another server with the flag set, do not fret.
Shutdown the node, and remove the
raft/ folder from the data directory. This
will remove the bad state caused by being in
-bootstrap mode. Then restart the
node and join the cluster normally.