I am using fileebat 6.5.1. We have server that produce large volume of data per day(75 gb).
I am facing memory issue on this server for filebeat. Filebeat is keep on increasing memory usage every day almost 7GB.
can you please suggest some configuration to overcome this issue.
there are number of process running on that servers. Each process keep on creating log file on average say in every hour by log rotation. For example when file is 5MB it will create rotation files. Almost producing 70 GB data per day. registerty file size is almost 14MB.
I have started filebeat yesterday (started with memory utilization 163MB)and today it reach upto 5703 MB. If this continue we will face disk running out of space.
Is the configuration is correct ? Can you suggest proper configuration to handle this issue.
Is there no other way other then moving to 6.8.1 ?
Can you please also advise that close_timeout: 5m is the correct configuration ? as per his description on fileebeat docs it will close the file handler after given interval irrespective of the reading position in file. We have applied this setting but still facing memory issue. Does this means it is not closing handler properly ?
This scenario sounds like you have a many many small files. About 1400 files per day.
This is quite a lot. Is there a particular reason to rotate at 5MB? How about switching to 1GB (have 70 files per day)?
The fixes in 6.8.1 should fix a memory leak exactly in this scenario: many many small files with rather short processing time.
For how long to you keep these files around? How many files do you have on disk?
14MB registry file is hughe. Sounds like loads of CPU time is spend when the registry file snapshot is written. Does the regsitry file hold files not on disk anymore (registry file is in JSON, you can easily parse it)?
A low close_timeout is ok. If the file is still on disk and the offset does not point to the end of file, then filebeat will pick it up again and continue from last known position. The memory leak found in filebeat is triggered by closing files often, though. Best way to mitigate this is: reduce number of files to process in total and make sure we don't need to close early.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.