Minimum hardware requirement for 500 million documents


I have a document with about 15 fields and I want to index 500million documents (around 350 GB of data) with elastisearch. I also will be querying this data with medium to complex queries including aggregations etc.

Can somebody please suggest the minimum hardware configuration that I should take into account for this ES server with one node?


You can easily store that amount of data on a single JVM and the max heap you want for a single JVM is 30GB.
Ideally you should test to see how your data structure applies to your use case and Elasticsearch.


Thanks for the reply.

Following are some of the follow up queries:

I tested with 25 million records, and found that I can query my data with 1GB of allocated heap.
But, if I increase the records to 30million and then query the data with 1GB of allocated heap, I get the following exception:"

     [FIELDDATA] Data too large, data for [bad_score] would be larger than limit of  
     [623326003/594.4mb]]; nested: CircuitBreakingException[[FIELDDATA] Data too large, data for 
     [----] would be larger than limit of [623326003/594.4mb]]; }
    at java.util.concurrent.ThreadPoolExecutor.runWorker(
    at java.util.concurrent.ThreadPoolExecutor$
  1. Please let me know in detail what are the different strategies to avoid this error in the Prod environment, Do we need to keep monitoring the elasticsearch nodes and then scale horizontally as and when required?

  2. Can I assume that if I increase my data n times of 25 million then I also need to increase my heap memory by n of 1GB?

  1. Look at doc values.
  2. It's more complicated than that, but hard to quantify. You need to test.
1 Like