I am trying to reduce the volume of my data in metrics monitoring. Currently, I filter out fields using Logstash and retain only the fields that I need for my dashboards.
I have 7 servers, and each server generates 204 MB of data per day. So, for 7 servers, that totals 1.4 GB per day.
Is there a way to further reduce the data? See the picture below; these are the fields I have retained for the dashboards.
From my understanding you already are using default tools and tried to delete as much fields as you could, now will come your indexing strategy and how much nodes do you have ?
Can you increase the metric frequency, let's say 1 log per minute instead of 30 ?
How much precision do you need to have ?
How much retention do you need to have ?
You can also trim much more metadata fields if you dont need them.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.