Data is duplicating without restart of any services of kibana

for payment URL index hits are coming after every one min for 2 different urls.
i.e for each url we are getting 60 hits per min.

we have created sincedb path for index and observed data but we found that data is duplicating sometimes without taking restart of kibana services
As attached in below screenshot we can see sudden rise in the graph.

Please find configuration file for index as below.

input {
file
{
path => ["/log/AnsibleLogs/Formatfile.txt"]
tags => ["paymenturls"]
# sincedb_path => "/dev/null"
# start_position => "end"
# ignore_older => "3m"
sincedb_path => "/var/lib/logstash/plugins/inputs/file/.sincedb_9e5a6659de603250b13193c70cd50ef7"

     }

}

filter{

if "paymenturls" in [tags]{

grok {
match => { "message" => "%{DATA:url}, %{DAY:day} %{MONTH:month} (?[0-9 ]{2}) %{TIME:time} %{WORD:Zone} %{YEAR:year}, %{NUMBER:Status_code1}, %{GREEDYDATA:ErrorMessage}" }
}

mutate {
add_field => {
"timestamp" => "%{date} %{month} %{year} %{time}"
}
}

}

}

output
{
if "paymenturls" in [tags]
{
elasticsearch
{
hosts => [ "https://elkmonp4.newindia.co.in:9200"]
index => "niapaymenturls-%{+YYYY.MM.dd}"
ssl => true
cacert => ["/etc/logstash/ELKMONP4.crt"]
user => "elastic"
password => "Password@123"
}
}
}

Is this a network share or a local path in your server?

Also, is this file rotated or recreated frequently?

Not a network shared file
This file is recreated based on the changes at the source server

The sincedb functionality assumes the file is appended to and not rewritten. If a file is rewritten with the same initial content and some new lines it will appear as a new file as the inode will change and the whole file will be reprocessed. This may explain why you are seeing duplicates.

1 Like

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.