i have 3 data node cluster. each has 15gb memory for ES and 15gb for lucen
have each record of sixe 5kb around around 100 fields
pumping documents at 5000 per seconds
after some time i am getting write timeout
is it too much to expect from this configuration or should i start looking into some parameters
I see mem and cpu usage is normal not too high
Io wait also seems to be okay... i dont see huge wait
any suggestions will be very helpfull
I have 2HDD, initially i had 48 threads writing to ES but reduced to 32 thinking this could be bottleneck... I had set batch size to 5000... initially for some hours will be able to write with in milliseconds and it gradually increases to seconds and then till 30sec and timeout happens
I don't have ant text search index enabled."index":"not_analyzed" for string fields
"doc_values": true,
"norms": { "enabled": false }
Have 2 indices.. pumping each at the rate 2500 msgs per second
Have set write timeout to 30s
ESMetricStore - onFailure : Error in bulk indexing org.elasticsearch.ElasticsearchTimeoutException: Timeout waiting for task. at org.elasticsearch.action.support.AdapterActionFuture.actionGet(AdapterActionFuture.java:70) ~[rtp-kafkastreams-1.0-SNAPSHOT-jar-with-dependencies.jar:na] at net.appcito.elasticsearch.util.ESMetricStore.executeBatch(ESMetricStore.java:375) [rtp-kafkastreams-1.0-SNAPSHOT-jar-with-dependencies.jar:na] at net.appcito.elasticsearch.util.ESMetricStore.execute(ESMetricStore.java:290) [rtp-kafkastreams-1.0-SNAPSHOT-jar-with-dependencies.jar:na] at net.appcito.elasticsearch.util.ESMetricStore$MetricStoreTask$ExecuteMetricStoreTask.run(ESMetricStore.java:736) [rtp-kafkastreams-1.0-SNAPSHOT-jar-with-dependencies.jar:na] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) [na:1.7.0_111] at java.util.concurrent.FutureTask.run(FutureTask.java:262) [na:1.7.0_111] at