I have a case that works fine on logstash 1.5.x but breaks in 2.2, 2.3, and 5.0 (I haven't tried 2.0 or 2.1). I have hourly syslog files that are approximately 3 gig each. I'm trying to backfill logs going back to January. Here's my config on my shipper:
@Janet This is really weird it should just be able to read everything, Any errors in the logs? You don't have any filters? If you run it with --debug do you get more details? Warning theses log can get a bit more noisy.
The plugin "logstash-input-file" has an argument called "ignore_older", by default, it will ignore the old file which was last modified before 24 hours. You can try to set it bigger to see whether it works well or not.
ignore_older
Value type is number
Default value is 86400
When the file input discovers a file that was last modified before the specified timespan in seconds, the file is ignored. After it’s discovery, if an ignored file is modified it is no longer ignored and any new data is read. The default is 24 hours.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.