Hello,
I am testing ELasticsearch to deploy it in production cluster. The first days it was working perfectly, and then I began having trouble when I try to search logs using Kibana, and the response was taking a long time, and now it's almost not working at all. I am getting Time out Errors
and then Kibana crash.
Here are some errors that I am getting in Kibana:
server log [09:24:33.353] [warning][kibana-monitoring][monitoring][monitoring][plugins] Unable to bulk upload the stats payload to the local cluster
server log [09:24:35.436] [error][plugins][taskManager] Failed to poll for work: Error: work has timed out
server log [09:24:37.741] [error][elasticsearch][taskManager] [TimeoutError]: Request timed out
server log [09:24:43.193] [warning][kibana-monitoring][monitoring][monitoring][plugins] Error: [cluster_block_exception] blocked by: [SERVICE_UNAVAILABLE/2/no master];
at respond (/home/ELK-8_0_0/kibana/node_modules/elasticsearch/src/lib/transport.js:349:15)
at checkRespForFailure (/home/ELK-8_0_0/kibana/node_modules/elasticsearch/src/lib/transport.js:306:7)
at HttpConnector.<anonymous> (/home/ELK-8_0_0/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:173:7)
at IncomingMessage.wrapper (/home/ELK-8_0_0/kibana/node_modules/lodash/lodash.js:4949:19)
at IncomingMessage.emit (events.js:203:15)
at endReadableNT (_stream_readable.js:1145:12)
at process._tickCallback (internal/process/next_tick.js:63:19)
server log [09:24:43.193] [warning][kibana-monitoring][monitoring][monitoring][plugins] Unable to bulk upload the stats payload to the local cluster
server log [09:24:52.008] [error][index][plugins][security][session] Failed to check if session index template exists: Request Timeout after 30000ms
Unhandled Promise rejection detected:
{ Error: Request Timeout after 30000ms
at /home/ELK-8_0_0/kibana/node_modules/elasticsearch/src/lib/transport.js:397:9
at Timeout.<anonymous> (/home/ELK-8_0_0/kibana/node_modules/elasticsearch/src/lib/transport.js:429:7)
at ontimeout (timers.js:436:11)
at tryOnTimeout (timers.js:300:5)
at listOnTimeout (timers.js:263:5)
at Timer.processTimers (timers.js:223:10)
status: undefined,
displayName: 'RequestTimeout',
message: 'Request Timeout after 30000ms',
body: false }
Terminating process...
server crashed with status code 1
And WArninig in an ELasticsearch node are :
[2020-11-16T09:39:36,599][WARN ][o.e.t.TransportService ] [MASTER-01] Received response for a request that has timed out, sent [25413ms] ago, timed out [15408ms] ago, action [internal:coordination/fault_detection/leader_check], node [{MASTER-03}{RCeMt0uXQie_ax_Sp22hLw}{ghlsACoWQKSva8mIlCKssQ}{X.X.X.X}{X.X.X.X:9300}{dilmrt}{ml.machine_memory=8365068288, ml.max_open_jobs=20, xpack.installed=true, data=hot, transform.node=true}], id [14671045]
[2020-11-16T09:39:36,600][WARN ][o.e.t.TransportService ] [MASTER-01] Received response for a request that has timed out, sent [14407ms] ago, timed out [4402ms] ago, action [internal:coordination/fault_detection/leader_check], node [{MASTER-03}{RCeMt0uXQie_ax_Sp22hLw}{ghlsACoWQKSva8mIlCKssQ}{X.X.X.X}{X.X.X.X:9300}{dilmrt}{ml.machine_memory=8365068288, ml.max_open_jobs=20, xpack.installed=true, data=hot, transform.node=true}], id [14671080]
[2020-11-16T09:39:37,824][WARN ][o.e.m.f.FsHealthService ] [MASTER-01] health check of [/var/lib/ELK-8_0_0] took [12006ms] which is above the warn threshold of [5s]
[2020-11-16T09:40:14,831][WARN ][o.e.t.InboundHandler ] [MASTER-01] handling inbound transport message [InboundMessage{Header{5844}{8.0.0}{10824701}{true}{false}{false}{false}{indices:data/write/bulk[s][r]}}] took [6203ms] which is above the warn threshold of [5000ms]
[2020-11-16T09:40:14,832][WARN ][o.e.t.InboundHandler ] [MASTER-01] handling inbound transport message [InboundMessage{Header{5645}{8.0.0}{42084006}{true}{false}{false}{false}{indices:data/read/search[phase/query]}}] took [6203ms] which is above the warn threshold of [5000ms]
Could you give me some advices to solve these errors.
Knowing that I have 5 Elasticsearch nodes: 3 master nodes and 2 data-nodes
and JVM is configured to 4G in each node
Thanks