I have a question about how to manage an ES cluster where the servers have
different sized data disks. I have 4 servers with 1 x 1TB disks for ES
data, and I am planning on adding an additional 2 servers that will have 4
x 4TB disks for ES data.
Is ES "aware" of the capacity of its data disks at all and do anything to
balance the load evenly, or will that fall to me to manually move shards
around to do that?
I have a question about how to manage an ES cluster where the servers have different sized data disks. I have 4 servers with 1 x 1TB disks for ES data, and I am planning on adding an additional 2 servers that will have 4 x 4TB disks for ES data.
Is ES "aware" of the capacity of its data disks at all and do anything to balance the load evenly, or will that fall to me to manually move shards around to do that?
I have a question about how to manage an ES cluster where the servers have
different sized data disks. I have 4 servers with 1 x 1TB disks for ES
data, and I am planning on adding an additional 2 servers that will have 4
x 4TB disks for ES data.
Is ES "aware" of the capacity of its data disks at all and do anything to
balance the load evenly, or will that fall to me to manually move shards
around to do that?
It'll make sure it doesn't fill up the disks if you enable the disk
allocaiton decider (enabled by default from 1.3 onwards) but beyond it
won't do anything like balance shards proportionally to disk capacity. In
many cases that wouldn't be the right thing to do anyway. You can use the
allocation awareness api stick shards to some machines or force them away
from others but there is no way to say "this node is twice as good put
twice as many shards on it" without running two copies of Elasticsearch on
the node.
I have a question about how to manage an ES cluster where the servers
have different sized data disks. I have 4 servers with 1 x 1TB disks for
ES data, and I am planning on adding an additional 2 servers that will have
4 x 4TB disks for ES data.
Is ES "aware" of the capacity of its data disks at all and do anything to
balance the load evenly, or will that fall to me to manually move shards
around to do that?
It'll make sure it doesn't fill up the disks if you enable the disk
allocaiton decider (enabled by default from 1.3 onwards) but beyond it
won't do anything like balance shards proportionally to disk capacity. In
many cases that wouldn't be the right thing to do anyway. You can use the
allocation awareness api stick shards to some machines or force them away
from others but there is no way to say "this node is twice as good put
twice as many shards on it" without running two copies of Elasticsearch on
the node.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.