Loading

Design for resilience

ECE ECK Elastic Cloud Hosted Self Managed

Distributed systems like Elasticsearch are designed to keep working even if some of their components have failed. As long as there are enough well-connected nodes to take over their responsibilities, an Elasticsearch cluster can continue operating normally if some of its nodes are unavailable or disconnected.

Elasticsearch implements high availability (HA) at three key levels:

  • Node level: Running multiple nodes within the cluster to avoid single points of failure and maintain operational stability.
  • Zone level: Distributing nodes and shards across availability zones to tolerate multi-node failures at infrastructure level and maintain service continuity.
  • Index level: Configuring shard replication to protect against data loss and improve search performance by distributing queries across multiple copies.

Each of these HA mechanisms contributes to Elasticsearch’s resilience and scalability. The appropriate strategy depends on factors such as data criticality, query patterns, and infrastructure constraints. It is up to you to determine the level of resiliency and high availability that best fits your use case. This section provides detailed guidance on designing a production-ready Elasticsearch deployment that balances availability, performance, and scalability.

Note

In the context of Elasticsearch deployments, an availability zone, or simply zone, represents an isolated failure domain within your infrastructure. Depending on the design of your cluster, this could be a physically separate data center, a different section within the same data center, distinct server racks, or logically separated node groups. The goal of using availability zones is to minimize the risk of a single point of failure affecting the entire deployment.

For example, in Elastic Cloud Hosted, availability zones correspond to the cloud provider’s availability zones. Each of these is typically a physically separate data center, ensuring redundancy and fault tolerance at the infrastructure level.

Learn more about nodes and shards, or view reference architectures for high availability deployments.

There is a limit to how small a resilient cluster can be. All Elasticsearch clusters require the following components to function:

A resilient cluster requires redundancy for every required cluster component. This means a resilient cluster must have the following components:

  • At least three master-eligible nodes
  • At least two nodes of each role
  • At least two copies of each shard (one primary and one or more replicas, unless the index is a searchable snapshot index)

A resilient cluster needs three master-eligible nodes so that if one of them fails then the remaining two still form a majority and can hold a successful election.

Similarly, redundancy of nodes of each role means that if a node for a particular role fails, another node can take on its responsibilities.

Finally, a resilient cluster should have at least two copies of each shard. If one copy fails then there should be another good copy to take over. Elasticsearch automatically rebuilds any failed shard copies on the remaining nodes in order to restore the cluster to full health after a failure.

Important

Failures temporarily reduce the total capacity of your cluster. After a failure, the cluster must also perform background activities to restore itself to health. You should make sure that your cluster has the capacity to handle your workload even if some nodes fail.

Depending on your needs and budget, an Elasticsearch cluster can consist of a single node, hundreds of nodes, or any number in between. When designing a smaller cluster, you should typically focus on making it resilient to single-node failures. Designers of larger clusters must also consider cases where multiple nodes fail at the same time.

The following pages give some recommendations for building resilient clusters of various sizes:

In addition, Resilience in Elastic Cloud Hosted and Elastic Cloud Enterprise deployments outlines how ECH and ECE orchestrators implement resilience, and offers guidance to ensure your deployments follow best practices.

When designing a resilient Elasticsearch cluster, avoid routing client traffic, whether from Kibana, Logstash, Beats, Elastic Agent, or other applications, to a single node. Relying on one node introduces a single point of failure, which can compromise data availability if that node becomes unavailable.

To avoid this, traffic should be distributed across multiple nodes using one of the following approaches:

  • Send traffic to all Elasticsearch nodes: A simple and effective strategy, especially in small clusters.
  • Send traffic to a dedicated subset of nodes: For example, route client requests only to hot nodes, or to coordinating-only nodes in clusters using data tiers.
  • Use an external load balancer or reverse proxy: This can manage routing and failover logic outside the cluster itself, and it's the default method in orchestrated platforms such as Elastic Cloud Hosted, Elastic Cloud Enterprise, and Elastic Cloud on Kubernetes.

Refer to resilience in small clusters and larger clusters for more details and recommendations.

Note

To ensure high availability in Kibana, configure it to send requests to multiple Elasticsearch nodes.