I know that I'm asking a rather generic question that can't be answered without understanding my specific use case. However, I'm looking for some lower bounds. What's the minimum number of nodes, size of the node (in terms of number of CPUs, RAM), that would be required for
N number of documents distributed across
S TB of data with
I'm not expecting a purely mathematical answer to this, but want to know if any prior research has been done to establish at least some correlation.