I ingested csv in elasticsearch using logstash and also put that ingesting script running in background for continues ingestion.
It's ingest if new row added to csv but if I have any row values update then it's not updating on elasticsearch, like logstash script not able to figure out that it's input csv updated.
ex. My csv content:
10/10/2017 10:22,Chicago,23412 10/10/2017 10:30,New York,67890
And it's already ingested in elasticsearch.
Now I update csv as:
10/10/2017 10:22,Chicago,56789 10/10/2017 10:30,New York,67890
So I have column 3rd value update in first row but after this update logstash script runing on background continuously not updating values in ealsticsearch.
Is there any parameters I have to add in my lofgstash conf.
My log stash conf as below:
input {
  file {
    path => "population.csv"
    start_position => "beginning"
    sincedb_path => "/dev/null"
    ignore_older => 0
  }
}
filter {
  csv {
    separator => ","
    columns => ["timestamp","city","population"]
  }
  date {
    match => [ "timestamp", "MM/dd/YY HH:mm"]
    target => "timestamp_ingestion"
  }
}
output {
   elasticsearch {
     hosts => "http://localhost:9200"
     index => "data-index"
  }
stdout {}
}
Can you please suggest any config parameters I need to add?