All: I searched this list and only found one message from 2015 related to this and it was not answered.
I have an elasticsearch 5.4 cluster. It has 20 data nodes with 5 of them also serving as master
nodes. Each node has two 512GB disks for data (/data01 and /data02).
The following is the settings of one node:
[oe_admin@oedevnode26 ~]$ curl -k --cert "/etc/pki/java/temp.pem" -s -XGET "https://oedevnode09:9201/_nodes?pretty&filter_path=nodes.*.settings.path.data"
{
"nodes" : {
"cr1kHpTGRQmZSj59gs9Bcw" : {
"settings" : {
"path" : {
"data" : [
"/data01/elasticsearch/oedl/data",
"/data02/elasticsearch/oedl/data"
]
}
}
},
The issue is that elastic is not balancing the data across both disks. So /data01 is filling up and exceeding 80% which is triggering a warning from my nagios monitoring software while /data02 has minimal data (ES shares these disks with hadoop hdfs). Here is what a typical DF looks like on a node:
/dev/sdd1 512G 74G 439G 15% /data02
/dev/sdc1 512G 408G 105G 80% /data01
My watermark settings are as follows:
[oe_admin@oedevnode26 ~]$ curl -k --cert "/etc/pki/java/temp.pem" -s -XGET "https://oedevnode09:9201/_cluster/settings?pretty=true"
{
"persistent" : { },
"transient" : {
"cluster" : {
"routing" : {
"allocation" : {
"disk" : {
"watermark" : {
"low" : "80%",
"high" : "75%"
}
}
}
}
}
}
}
I have tried the default settings and the goal of these was to get ES to move data to the second disk.
It appears that the watermark settings are being applied to the combined disk space versus each individual disk.
Is there are way to force ES to balance the space used across both disks?
Any easy ways to even balance things manually?
If I switch the order of the disks in the path will that change the order that ES writes to the disks?
Sorry if a missed any post already here that answers this problem.