hello,
I have configured the following s3 input in logstash, and it keeps reading the same file from the given S3 bucket. I noticed that the given sincedb file ${DATA_DIR}/.sincedb_activities.txt is being used and it contains the date and time corresponding to the last modified value of the file on S3.
What am I missing? Do I need to set the parameter delete to true ? As far as I understand, that should not be necessary, as the sincedb mechanism should guarantee files are processed only once. Thanks
#############
# INPUT #
#############
input {
s3 {
access_key_id => "${AWS_ACCESS_KEY_ID}"
secret_access_key => "${AWS_SECRET_ACCESS_KEY}"
endpoint => "${S3_ENDPOINT}"
bucket => "${INPUT_BUCKET}"
temporary_directory => "/tmp/logstash"
prefix => "activities"
additional_settings => {
force_path_style => true
follow_redirects => false
}
codec => multiline {
charset => "UTF-8"
auto_flush_interval => 5
pattern => '^([0-9]+|id_activ),'
negate => true
what => "previous"
}
sincedb_path => "${DATA_DIR}/.sincedb_activities.txt"
add_field => {
"type_fichier" => "db_activity"
"[@metadata][index]" => "${ACT_IDX}"
"[@metadata][pipeline]" => "ACT"
}
}
}