Hi,
I have a cluster with almost all nodes 111 shards and only 2 nodes 108.
I understand that ES balance the cluster with shard count.
cluster has around 19 nodes, 3 master eligible of them 2 data as well and 16 data nodes.
these two nodes with less shards are always reaching low watermark. Even though the shards are around same size.
I have few nodes with 600gb free and these two nodes with less shard count have only 70gb free. This is because of the few indices with very less size and few with more size.
ES version: 7.5.2
Questions:
How to move replica shards from one node to another. (delete and direct to allocate them on a specific node or is there any direct way to issue a command using reroute).
When rebalancing the cluster, how ES calculates what shard to move from one node to another.
Hi @DavidTurner
It is normal if its temporary.
But those two nodes are always in a low watermark level,
I am guessing this is because all the indexing always happening to these nodes because of the lesser shard count. (even if the rebalancing happening, because of the indexing to these nodes, they are reaching again the watermark levels).
Even If I create a no data index with multiple shards on that node only to increase the shard count, cluster is rebalancing (moving away other indices) and somehow still after some time these nodes still with full capacity and less shards in all.
That is why I was trying to see how ES selects a shard to move in order to rebalance.
MY ES version is 7.5
It's normal to temporarily exceed the high watermark, but it's also completely normal to exceed the low watermark on some nodes basically forever. This doesn't need any action.
The docs to which I linked earlier explain this, and the section just above explains how rebalancing works.
Ah I see that you're on an older version, which means it will be emitting a lot of misleading logs about watermarks. This was fixed in more recent versions:
I wouldn't have thought so, no. This is also covered in the reference I linked above:
A cluster with equal disk usage on every node typically performs no better than one that has unequal disk usage, as long as no disk is too full.
"Too full" is definitely above the high watermark (pretty much by definition: if you want to keep more space free for performance reasons then you should adjust the high watermark to suit)
This is also covered in the reference I linked above:
If a node exceeds the high watermark then Elasticsearch will solve this by moving some of its shards onto other nodes in the cluster.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.