I am unexperienced on Elastic ,I have used only twice on some of my side projects. But my company uses it on AWS Opensearch. I am a cloud engineer ,therefore my task has been to check costs ,and optimize.
I came upon an Opensearch Cluster on production ,that has 26 nodes size : r5.large.search :: 2Vcpu , 16Gb Ram.
I know for sure that there is a problem with either the setup or how applications use this cluster, because I know we do not have that much data to process, index or search.
The cpu stays on each 3% , and Memory stays 97% on each Node.
Now I have attached a screen shot that shows the nodes.
I don't have any plans to go into the depths of opensearch or elasticsearch.
But I know something is wrong because we do not have that much data to work with. But I thought someone might see the metrics and say something is off.
About my role , I just want to see where the money is going and if there is a way to optimise our setup. I like to automate with as little money as possible.
Remove a node every couple of days from the cluster. Until someone notices that something “seems slower”.
Then stop.
You’ll have optimised AND saved money.
Suggestion is , slightly, in jest.
See if you can find someone involved in provisioning the cluster , and ask them how they arrived at the size of your cluster. There’s likely an architect involved. Maybe even a design document with some numbers. Compare with reality.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.