Attempting to remove a node from a 5.6 cluster and did:
# cat /tmp/rmnode
{
"transient" :{
"cluster.routing.allocation.exclude._ip" : "10.3.3.12"
}
}
# curl -s -XPUT 10.3.3.12:9200/_cluster/settings -H 'Content-Type: application/json' -d @/tmp/rmnode | jq .; echo
{
"acknowledged": true,
"persistent": {},
"transient": {
"cluster": {
"routing": {
"allocation": {
"exclude": {
"_ip": "10.3.3.12"
}
}
}
}
}
}
reallocation also started but it now seems to have stop and still have 15 shards on node:
# curl -s -XGET 10.3.3.12:9200/_cluster/health?pretty | jq .
{
"cluster_name": "mx9es",
"status": "green",
"timed_out": false,
"number_of_nodes": 5,
"number_of_data_nodes": 4,
"active_primary_shards": 140,
"active_shards": 280,
"relocating_shards": 0,
"initializing_shards": 0,
"unassigned_shards": 0,
"delayed_unassigned_shards": 0,
"number_of_pending_tasks": 0,
"number_of_in_flight_fetch": 0,
"task_max_waiting_in_queue_millis": 0,
"active_shards_percent_as_number": 100
}
# curl -s -XGET '10.3.3.12:9200/_cat/allocation?v'
shards disk.indices disk.used disk.avail disk.total disk.percent host ip node
88 241.8gb 242.6gb 202.6gb 445.2gb 54 10.3.3.11 10.3.3.11 cbdB
15 38.3gb 39gb 406.2gb 445.2gb 8 10.3.3.12 10.3.3.12 cbdC
89 249.2gb 250.2gb 634.2gb 884.4gb 28 10.3.3.13 10.3.3.13 cbdD
88 272gb 273.2gb 175.8gb 449.1gb 60 10.3.3.10 10.3.3.10 cbdA
How to get the last 15 shards to move elsewhere?