There is no such thing as a single excel sheet where you type in the number of documents and get back the throughput. Everything is an approximation because of many different factors like document, mapping complexity, queries per second, query complexity, etc...
I would advise you to go ahead and benchmark your cluster with your own data. You can start simple with a single node, having a single shard and add data to it while querying it. See if it holds up your SLAs reagrding reading and writing. At some point you found out the sweet spot per shard and can start scaling based on nodes.
There is also a benchmarking tool for Elasticsearch called rally, you should take a look at.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.