Can't change timestamp of logstash to timestamp of log

Can't replace timestamp of logstash with timestamp of log file csv and send to elasticsearch.
in debugger all work.
When I change output to elasticsearch it's write that is work but after I create index patern in kibana I don't see something in discovery.
If I delete date filter all work correctly but I need change timestamp.
version of Elastic Stack 7.10.1
Java openjdk 14 2020-03-17
configuration file of logstash

input {
	file {
		path => "C:/Users/vladi/Desktop/Elastic_Stack/*Z.csv"
		start_position => "beginning"
		sincedb_path => "NULL"
	}
}
filter {
	csv {
		separator => ","
		skip_header => "true"
		columns => [ "timestamp","number" ]
		remove_field => [ "path", "host", "@version"]
	}
	mutate{
		convert => { "number" => "integer" }
	}	
	date {
		match => ["timestamp", "YYYY-MM-dd'T'HH'_'mm'_'ss.SSS'Z'"]
		timezone => "UTC"
		add_field => { "debug" => "timestampMatched"}
		remove_field => [ "timestamp" ]
	}
}
output {
	stdout { codec => rubydebug }
}

I get:

{
        "number" => 1232,
       "message" => "2020-12-10T09_24_13.986Z,1232\r",
    "@timestamp" => 2020-12-10T09:24:13.986Z,
         "debug" => "timestampMatched"
}

When I change output to elasticsearch after creating index patern and entering to discovery I take error:


How can I solve it?

You have kibana configured to show data from "This week", that would definitely not include December 10th from last year.

Thanks

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.