I was referring to following for sizing the shards,
I referred the following sections in the article,
#Aim for shard sizes between 10GB and 50GB
#Aim for 20 shards or fewer per GB of heap memory
I wanted to understand if we can scale Single Elasticsearch Node up to ~29.29 TB using the following configuration?
• Max JVM Heap size - 30 GB
• Shard per GB - 20
• Total Shards – 600
• Shard size – 50 GB
If possible, what all things needs to be taken into consideration as part effective performance management.
The numbers given are guidelines for maximum numbers. Each shard requires some overhead in terms of heap and you also need to have enough spare heap space to handle indexing requests and queries. Exactly how much data a node can handle will depend on your data, mappings, shard sizes and usage patterns. There is no guarantee that a node with 30GB heap can handle 600 shards of 50GB each.
Also note that query performance is likely to be affected by the amount of data each nodes hold and the type of storage used. If you have requirements on max query latency this may be what is limiting the data density.
Any ballpark number how much single node can scale with above configuration and when to go for a new node?
It depends on your data and how heavily you index and query so I am afraid you will need to test/benchmark.
This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.