Hello,I have a problem about my cluster's master node heap cost!
after some days ,the heap of master node cost chart is this:
after i restart master node,heap used chart is this:
every day ,i have job to remove ,close,force merge indices(daily indices),so es data size is stable,but the heap used bigger and bigger , i have to adjust heap size or restart master node(dedicated master node)
my es version is 5.0.1, i guess is there a memory leak problem?
Hmm, could you create a heap dump of your master node when it runs out of memory next time to see how the heap memory is used? You can analyze it yourself or email it to me, just don't post it here since it might contain sensitive information about your cluster setup.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.