I am using logstash to read .gz files. The gzip files are around 10 to 15GB for which logstash was unable to process and crashing. So I have split the .gz files using 7zip split based on size of 3GB .gz file chunks. i see logstash easily processing so far except the issue that it cannot determine the EOF which is significant in read mode to process .gz files. and logstash keep processing the same file and not reading the other files specified in path.
This error means your .gz is corrupted or write stream didn't close it yet. We may skip such cases when error happens but you may miss the data to the outputstream. So, suggestion will be to fix the corrupted file or waif till processing file is closed.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.