﻿---
title: Merge settings
description: A shard in Elasticsearch is a Lucene index, and a Lucene index is broken down into segments. Segments are internal storage elements in the index where...
url: https://www.elastic.co/elastic/docs-builder/docs/3028/reference/elasticsearch/index-settings/merge
products:
  - Elasticsearch
applies_to:
  - Elastic Stack: Generally available
---

# Merge settings
<admonition title="Settings supported in Serverless">
  Elastic Cloud Serverless projects [restrict the available Elasticsearch settings](https://docs-v3-preview.elastic.dev/elastic/docs-builder/docs/3028/deploy-manage/deploy/elastic-cloud/differences-from-other-elasticsearch-offerings#elasticsearch-differences-serverless-settings-availability) to a supported subset, identified with a `Serverless` badge next to the setting name. For a complete list of available index settings, refer to the [Serverless index settings list](https://www.elastic.co/elastic/docs-builder/docs/3028/reference/elasticsearch/index-settings/serverless).
</admonition>

A shard in Elasticsearch is a Lucene index, and a Lucene index is broken down into segments. Segments are internal storage elements in the index where the index data is stored, and are immutable. Smaller segments are periodically merged into larger segments to keep the index size at bay and to expunge deletes.
The merge process uses auto-throttling to balance the use of hardware resources between merging and other activities like search.

## Merge scheduling

The merge scheduler controls the execution of merge operations when they are needed.
Merges run on the dedicated `merge` thread pool.
Smaller merges are prioritized over larger ones, across all shards on the node.
Merges are disk IO throttled so that bursts, while merging activity is otherwise low, are smoothed out in order to not impact indexing throughput.
There is no limit on the number of merges that can be enqueued for execution on the thread pool.
However, beyond a certain per-shard limit, after merging is completely disk IO un-throttled, indexing for the shard will itself be throttled until merging catches up.
The available disk space is periodically monitored, such that no new merge tasks are scheduled for execution when the available disk space is low.
This is in order to prevent that the temporary disk space, which is required while merges are executed, completely fills up the disk space on the node.
The merge scheduler supports the following *dynamic* settings:
<definitions>
  <definition term="index.merge.scheduler.max_thread_count">
    The maximum number of threads on a **single** shard that may be merging at once. Defaults to `Math.max(1, Math.min(4, <<node.processors, node.processors>> / 2))` which works well for a good solid-state-disk (SSD). If your index is on spinning platter drives instead, decrease this to 1.
  </definition>
  <definition term="indices.merge.disk.check_interval">
    The time interval for checking the available disk space. Defaults to `5s`.
  </definition>
  <definition term="indices.merge.disk.watermark.high">
    Controls the disk usage watermark, which defaults to `95%`, beyond which no merge tasks can start execution.
    The disk usage tally includes the estimated temporary disk space still required by all the currently executing merge tasks.
    Any merge task scheduled *before* the limit is reached continues execution, even if the limit is exceeded while executing
    (merge tasks are not aborted).
  </definition>
  <definition term="indices.merge.disk.watermark.high.max_headroom">
    Controls the max headroom for the merge disk usage watermark, in case it is specified as percentage or ratio values.
    Defaults to `100GB` when `indices.merge.disk.watermark.high` is not explicitly set.
    This caps the amount of free disk space before merge scheduling is blocked.
  </definition>
</definitions>


## Merge multi-threading

In general, merges are completed concurrently using multiple threads. Generally, each individual merge is single-threaded, but multiple merges can be run in parallel.
There are some exceptions and certain merges can utilize multiple threads within the merging operation itself. These threads will be assigned from the same pool as the background merge threads.
The following *dynamic* settings control multithreaded merging:

<definitions>
  <definition term="index.merge.intra_merge_parallelism_enabled Elastic Stack: Preview">
    Enables or disables intra-merge parallelism for data types that support it (e.g. `dense_vector`). Defaults to `false`.
  </definition>
</definitions>