In general, memory is never wasted. If you are doing heavy faceting, geo
or sorting, memory will probably be your limiting factor. And even when
you aren't faceting, more memory will mean more segments of your index are
cached into the file system cache (which means more diskless operations).
It's hard to make a blanket statement about CPU and disk I/O. More is
always better, but how much you need is impossible to determine without
testing. CPU is used heavily by some search types, as well as during
segment merges while indexing. Disk I/O is a bottleneck during heavy
indexing, shard relocation and potentially some searches/facets if you hit
all the documents at once.
With regard to disk, if you RAID, use RAID in a performance mode (striping)
rather than for availability....you don't need HA because you can use
replicas. If you can get SSDs, they are a huge boost in performance over
Lastly, "medium" to "large" boxes tend to work better than "small" boxes.
Those are quoted for a reason, because it's hard to say what is "large"
for a given context. But it tends to be more economical to start with
medium/large machines and then start scaling out.
On Tuesday, September 17, 2013 8:14:21 AM UTC-4, Julien Naour wrote:
Is there some hardware recomendations for ES?
It seems that it's very difficult to evaluate and really dependant of the
But is there some well-known bottle necks or typical configuration? I/O or
My use case is pretty facets intensive.
You received this message because you are subscribed to the Google Groups "elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email to email@example.com.
For more options, visit https://groups.google.com/groups/opt_out.