What size heap are you using?
We are using Elasticsearch 2.1 after recently upgrading from 1.7. We are using 128GB heaps because we used to get Java out-of-memory errors. Now what we're on 2.1, we are trying to resolve a timeout problem we've been experiencing. It was recommended that we go from 128GB to 30GB for heap.
OK, I'm willing to give that a go.
Aren't we about to switch one problem for another? What is the implication of going to a smaller heap given we had so many out-of-memory issues before. Are there OOM java issues which have been resolved since ES 1.7?