We also tried to set a size limit for the fielddata cache (40%) but still getting the OOM exception on this request.
Some tips about our cluster topology:
5 data nodes
max heap : 8gb
286 indices
3072 shards
2,857,660,643 docs
4.07TB
elasticsearch node log with exception :
detailed query :
We would like to know how to make sure this kind of requests will not be able to crash our entire cluster and how to go further in the root cause analysis.
Which version of Elasticsearch are you running? If you are running a version prior to 5.4.2 you may be running into these issues: #25010 and #24941. There are also still some known issues around aggregations and OOM which we are tracking in #26012
An average of >10 shards per index and 5 data nodes? Having more shards than data nodes is useful if you plan on expanding out into more data nodes in future but otherwise it's a less efficient way to store the data.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.