Hi - I have a very simple index that I want to optimise for search speed. It is a 23gb read-only index - queries are randomly distributed across all documents (i.e. there are no real hot spots) and with the vast majority being standard match searches with filters and the odd geo query (bounding box). I am looking at the best way to fully optimise this for near real-time search. Presumably "index.store.type": "memory" will give a big uplift? Also, I know that a general rule of thumb is to allocate 50% of available RAM to ES and leave 50% for the OS but would this still be the case for a read-only index?
I am looking at AWS r3 instance types which give the most memory for $. There is an r3.xlarge with 31.5gb RAM or an r3.2xlarge with 61gb RAM. Based on the configuration above and the index size of 23gb which is unlikely to grow would there be much advantage going for the r3.2xlarge - they are double the price so if I can squeeze on the smaller instance that would be awesome.
Thanks in advance for any advice - I have been trawling the forums and SO but not found anything with quite the same requirements.