Unallocated shards when a node is removed

ES 2.1
AWS EC2, Private subnets.

When ever I remove a node, shards that belong to this node don't get re balanced

I think some times resetting the number of replicas solves the issue.

Most of the times I end up restarting nodes.

This is happening consistently every time.
On the master node I see this

delaying recovery of [7bs2em65onshoizh][1] as it is not listed as assigned to target node {Impala}{HXQsJaPeTP2DTVY7WzA_4Q}

any ideas why this might be happening?

Do you have allocation disabled?
Check _cluster/settings

allocation is not disabled

_cluster/settings shows

_cat/shards shows UNASSIGNED shards untill I restart one of the remaining nodes

More info

delayed_timeout is default 1m

there is no diskspace watermark limit on any of the remaining nodes.



Moving the discussion from https://github.com/alicegoldfuss/shardnado/issues/1#issuecomment-220253473

I am not sure what this is

I think you should check the settings on your indices and cluster, use the explain feature on the _reroute API and first find out why before you fix the why. :slight_smile:

btw, I didn't use this shardnado tool, I was just saying that some times ES doesn't allocate shards on its own.

our cluster is a basic one, doesn't have any routing values setup or anything.
As posted in earlier comments in this thread, cluster settings are empty. I haven't disabled allocation.

I checked on of the index settings also /index-name/_settings, nothing useful there about shards.

I can reproduce this, happens every time I remove a node.

please let me know how I can use this _reroute to debug this further.

hey the _reroute API allows you to manually reroute shards to a node. you can use a parameter called explain=true that would give you the reasons why this allocation could or could not be applied. That should tell you why shards are not allocated and / or if they are throttled. If you call that API with an empty body you can trigger a new round of rerouting and get some information about all the shards. That should give you a much better idea of what is going on. If you can paste that output here I can take a look. Also send me the index settings of the index that is not allocating.


1 Like

Ok, so I have run this command

curl -XPOST 'localhost:9200/_cluster/reroute?pretty&explain' -d '{
    "commands" : [ 
          "allocate" : {
              "index" : "1qp6axstrjub7ouw", "shard" : 1, "node" : "Nimrod"

this allocated the replica 1 to the node Nimrod which was unassigned before. Remaining all unassigned shards also got allocated after this.

here is the output of that command in dry_run mode if that is of any use

Can I see the index settings of that index 1qp6axstrjub7ouw

here is the output of that command in dry_run mode if that is of any use

I think this is a bug in delayed allocation that misses to kick off another round of shard allocation. is there a chance for you to upgrade to 2.3 at some point?
Also you can simulate that missing round of allocaiotn by calling _reroute with an empty body.

here is the output of cluster settings

[root@ip-172-31-48-137 ~]# curl localhost:9200/1qp6axstrjub7ouw/_settings?pretty
  "1qp6axstrjub7ouw" : {
    "settings" : {
      "index" : {
        "creation_date" : "1446015175417",
        "number_of_shards" : "5",
        "number_of_replicas" : "1",
        "uuid" : "gOZtw-IGS2ap2Cz6tJt4MA",
        "version" : {
          "created" : "2000051"

We will eventually move to 2.3 (may be in few weeks or 1-2 months not sure), I saw there are some breaking changes from 2.1-2.2, need to look into them and plan.

ok so can you provoke this problem again and see if an empty reroute fixes it?

curl -XPOST 'localhost:9200/_cluster/reroute' this does allocate those unassigned ones

perfect, I think you ran into one of those bugs where delayed allocation missed a reroute. Can you please upgrade to the latest and see if the bug persists? If so please open an issue on our issue tracker! thanks!