I have multiple servers running filebeat sending logs to two logstash servers (all in AWS), after a couple of days the logstash servers stop ingesting and report "Too many files open". An lsof shows thousands of connections from the filebeat servers stuck in CLOSE_WAIT.
plugin version for logstash-input-beats (5.0.6)
We have done some test, try to use hostname IP or a FQDN but the issue is the same if elastic data node go down logstash remain in closed wait connection.
Version rpm 5.6.6-1 (logstash-5.6.6-1.noarch)
Operating System Centos 7.4 x64
Config logstash.yml
# Settings file in YAML
#
Is your behavior the same as stated above, where you will eventually run out of file descriptors due to thousands of connections in a CLOSE_WAIT state? Does Logstash recover once the elastic data node recovers?
The only way for restore full functionality is restart logstash service.
The CLOSE_WAIT connection is only one for data node (one to one)
In our test we have 2 elastic data server if we stop one data node, logstash remain in close wait on http port for data node that we have shutdown, if we stop other elastic data server, logstash remain in close wait on second data node too, for restore connection we have to restart logstash
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.