We're currently in process of choosing new hardware for our elasticsearch cluster.
Our current cluster consists of 32 nodes and holds 42TB of data across 2000 indices.
We're choosing hardware from the specs our provider has.
One of options we're considering is the one that has 256GB RAM and 4x2TB NVMe SSD.
We're planning to join those in RAID0 which would get us 8TB NVMe SSD per node.
My questing is is it maybe a bit too much since some of our shards are pretty small and we may cross 20 shards or fewer per GB of heap memory boundary.
And seeing is this node has too much RAM as well (although it can be used as cache) we were considering splitting those nodes into 4 LXC containers 64GB RAM and 2TB each. Which option would be preferable from elasticsearch perspective?
why use raid0. you should use single disk /data01, /data02, /data03 etc... and elasticsearch will manage them. if you loose one disk you are only loosing 25% of shard on that node. if you use 8TB with Raid0 then one dead disk and you have 100% shard lost for that node.
256 RAM might be overkill and on that sense your logic is write to split it.
I am also in process of setting up same amount of NVME but 98gig ram 20 node cluster.
@elasticforme Your questing mentions VMs which implies performance overhead much larger then one of LXC (which is not VMs, but Containers), so I'm not sure if reply to your question is applicable here.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.