How did you arrive at those numbers? Did you do any tests or run any benchmarks? What is the hardware specification you plan to deploy these node types on? How many users will you have? What are the query latency requirements?
I would recommend having a look at the following resources:
You have not answered any of my questions from the previous post.
Have you run a test to come up with these numbers? If so, did you index at least a few GB? Did you optimize your mappings?
Based on the calculation I take it this is an average EPS. If so, what is the expected peak rate the cluster need to be able to keep up with?
How did you determine that 10 days is the optimal period to keep on the hot nodes. Is this due to query requirements? How did you determine that 8 hot nodes is sufficient?
What is the expected specification of the hot nodes? What type of hardware and storage will you be using?
What is the expected specification of the warm nodes? What type of hardware and storage will you be using?
Here I am looking to see what your users expect and how they will use the cluster. When sizing a cluster it is important to leave enough headroom for querying and not just size the cluster based on the maximum it can index.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.