I'd like to inquire about logstash limits. I have an event coming in to logstash with about 6700 records however it doesn't process any of these and ignore it.
Is there any configuration I can adjust in order for this to be processed?
We tried adjusting our JVM Heap to 9Gb because we noticed that the heap usage is constantly reaching the allocated limit (as seen on attached image)
We also increased our logstash pipeline workers to 20 (logstash.yml) even with our current resources:
CPU(s): 16
Core(s) per socket: 1
Socket(s): 16
MemTotal: 16249840 kB
But we still cannot see the data into logstash. Hoping for your advise.
What kind of data is it? What is the average size of a document? What does your config look like? Is there any errors that n the logs of Logstash or Elasticsearch (assuming you are sending data there)?
Also not that we're getting data with the same structure in other environments. however, the one in this particular production environment doesn't process data with a large amount of records.
Average size of document is about 1.39 KB. We got this number in the index management of Kibana by dividing the storage size to the total number of documents (1.2 MB / 865)
Logstash configuration is pretty long and can't be saved here. Is there anyway I can send that to you?
And yes, we are sending data to ES. No Errors on Logstash and ES logs.
The event-message went through with 1,600 records. So there really seems to be a limit on data volume logstash can process? Is it something we can configure? Appreciate your response.
You are setting index name and document id based on fields. Are these always present in the data? Could there be any issue with these resulting in invalid index names or document ids? I would recommend increasing the logging level in Logstash to DEBUG to see if you see something.
Yes. These are always present as the fields are primary keys in the database.
Also, we've set the logs to DEBUG since the start but we can't find anything related to our transmission index to fail processing the data. logstash-plain.log
Any ideas what might be causing this? It seems odd that logstash doesn't process 2000 to 6000 records but process only around 1600. Here's our yaml file as well logstash.yml
I had another quick look at your config, and suspect you might have a few issues with it. As far as I can see you are dropping the message field for events where [check][name] is keepalive early on, but I do not see this explicitly caught in the section where document ids are created (although I do not know what the data looks like so it could have one of the other fields). If this goes to the default fingerprint id generation, all documents will get the same ID as the message field does not exist for these records.
An easy way to check this would be to look at the index statistics for deletes, as that can be an indication of updates being performed. You could also disable setting your own ID in the elasticsearch output and see if this makes a difference.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.