I've been given a go ahead to get some training which is great but first I need to get the cluster up to the version the training it all for.
I've had some issues when I tried a test upgrade before but since then I've refreshed the hardware in the cluster and our config is a little different now. Previously we had a RAID 0 array and everything lived in the default paths. For the rebuild we got SSDs so I have the OS on a RAID 1 array with a partition for logging on there too and then the data lives on 2 SSDs mounted as /data1 and /data2
When I run the pre-upgrade checks on the cluster now I get
Default path settings are removed This issue must be resolved to upgrade. Read Documentation Details: nodes with settings: [node3.mydomain.local]
It doesn't seem to flag issues with node 1 or node 2 only node 3, they were all built from the same scripted install 2 weeks ago so they all have the same configs and if I open up the elasticsearch.yml on each I can see
# Path to directory where to store the data (separate multiple locations by comma): # path.data: [/data1/elasticsearch, /data2/elasticsearch] # # Path to log files: # path.logs: /logs/elasticsearch
from what I read here in the referenced documentation
that would suggest that all I need to do is specify path.data and path.logs which I have so everything is set as it should be but the health check says I need to fix something on just that node before I can continue.
any insights appreceated