Hi All,
We run logstash on multiple EC2 instances behind a loadbalancer for reliability purposes. We are thinking of using the S3 input plugin. Since the servers are created by auto-scaling process of AWS, they are exactly same.
I am trying to get some clarity around the behaviour of the S3 input plugin when multiple instances of Logstash are running polling the same S3 bucket and prefix.
My understanding is that they should be fine since each S3 object is key looks like a filepath but is not actually a filepath.
Sample code for my case will be like this. I delete the S3 object after reading it. So no need to track the last handled file.
input
{
s3
{
bucket => "testbucket"
prefix => "get/this/data"
region => "us-east-2"
delete => true
interval => 100
sincedb_path => "/dev/null"
additional_settings => {
"force_path_style" => true
"follow_redirects" => false
}
}
}