We have a 10 node elasticsearch cluster which is receieving roughly 10k/s
worth of logs lines from our application.
Each elasticsearch node has 132gb of memory - 48gb heap size, the disk
subsystem is not great, but it seems to be keeping up. (This could be an
issue, but i'm not sure that it is)
The logs path is:
app server -> redis (via logstash) -> logstash filters (3 dedicated boxes)
We currently bulk import from logstash at 5k documents per flush to keep up
with the volume of data that comes in.
Here are the es non standard configs.
Field Data cache tuning
#Segment Merging Tuning
We have not had this cluster stay up for more than a week, but it also
seems to crash for no real reason.
It seems like one node starts having issues and then it takes the entire
Does anyone from the community have any experience with this kind of setup?
Thanks in Advance,
You received this message because you are subscribed to the Google Groups "elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email to email@example.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/d04a643e-990b-40b0-b230-2ba560f08eea%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.