Master not discovered exception

Hi there,

We had 4 nodes and added fifth one yesterday. We were expecting to find master automatically but got master_not_discovered_exception. Furthermore other 4 nodes act like there is no fifth node. All of nodes have same configuration.

     discovery: {
    zen: {
    ping: {
    unicast: {
    hosts: [
    "x.x.x.1",
    "x.x.x.2",
    "x.x.x.3",
    "x.x.x.4",
    "x.x.x.5"
    ]
    }
    }
    }
    },
    client: {
    type: "node"
    },
    http: {
    type: {
    default: "netty4"
    },
    port: "9200"
    },
    transport: {
    type: {
    default: "netty4"
    }
    }

What's the error?
What version?

Version 5.2.2
telnet to master ok. ping ok. But got the error below when i try to search on index

And when i check for the node information on master here i got

By the way all 4 nodes have this configuration

Please don't post pictures of text, they are difficult to read and some people may not be even able to see them.

Ok. Sorry about that Mark, Here is the answer without pictures.
Version 5.2.2
telnet to master ok. ping ok. But got the error below when i try to search on index

{
  error: {
    root_cause: [
      {
        type: "master_not_discovered_exception",
        reason: null
      }
    ],
    type: "master_not_discovered_exception",
    reason: null
  },
  status: 503
}

And when i check for the node information on master here i got

  {
    _nodes: {
      total: 4,
      successful: 4,
      failed: 0
    }
  }

By the way all 4 nodes have this configuration

 roles: [
    "master",
    "data",
   "ingest"
 ],

And the exception in the log

[2017-04-12T12:05:14,155][INFO][o.e.d.z.ZenDiscovery] [node-5] failed to send join request to master [{node-1}{rzzzzzzzzzzzzzzzzzzzzzzzzzz}{3aaaaaaaaaaaaaaaaaaaaa}{y.y.y.y}{y.y.y.y:9300}], reason [RemoteTransportException[[node-1][y.y.y.y:9300][internal:discovery/zen/join]]; nested: ConnectTransportException[[node-5][x.x.x.x:9300] connect_timeout[30s]]; nested: IOException[Connectiontimed out: no further information: x.x.x.x/x.x.x.x:9300];]

1 Like

Hi if you found an answer I will be happy to hear :slight_smile:

Hey Tomer,

It was because of Windows firewall. I added 9300 port to firewall on fifth node and its gone live.

1 Like

Hi

My case is similar but no issues with firewall. Here are details...

-Regards, dp

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.