We have 7 nodes cluster with 3 masters and 4 data nodes. We have one index with,
4 shards,
1 replica
size - 3.5GB
12894 documents
which is getting 100/s search requests normally and indexing rate is very low. But when peak hours reached index is hitting around 500 - 600 /s search requests. When that happens JVM heap is going up and hitting 99% and then data node is going down.
If that doesn't help increasing the heap size of your data node JVMs might be unavoidable to make 6.8.2 perform in this scenario. Another option though would be to upgrade to 7.x which has strong circuit breaker logic in place through the real-memory circuit breaker.
Thanks for quick response. I'll have a look on those areas. Meantime if i reduce no of shard to 1 and increase replica to 2 or 3 since index size only around 3.5GB with less documents, would that help?
If your shard size is indeed that small, yes moving to a single shard may be a good optimisation (especially when the indexing rate is low).
You'd have to use 3 replicas though to make full use of your 4 data nodes if I understand your setup correctly (with 2 one of your nodes will just be unused).
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.