we should implement really big data scenario with this features:
- input data is 1 million logs per second
- each log has almost 100 Bytes size
- we should retain data for 10 days
- system has 2,3 active users and they run maybe 1000 queries per day so our scenario is definitely heavy write
with this assumptions ( so many writes and small number of reads) , how much data each node in cluster should store ?