converting more than one column from string > integer completely invalidates the logstash output.
The values are interpreted as keys somehow
What am I missing when converting?
regardless of:
CSV { convert => {}}
mutate{ convert => {}}
converting more than one column from string > integer completely invalidates the logstash output.
The values are interpreted as keys somehow
What am I missing when converting?
regardless of:
CSV { convert => {}}
mutate{ convert => {}}
What do you mean by this?
Thank you for your reply.
I am quite new to the logstash pipeline, which means I am struggling to identify causes and problems:
my scenario is as follows:
i have +.csv s that are read via filebeat, and parsed via filter csv plugin.
csv {
autodetect_column_names => true
skip_header => true
separator =>";"
convert => {
"key1"=>"integer"
"key2"=>"integer"
...
}
When converting any or a number a colums, the output gets strange:
the original strings
"key1":"1"
"key2":"2"
that should look like
"key1":1
"key2":2
may become
"1":2
or
"key1":1
or
column2:"2"
I have a single worker and I (maybe) depending on the amount of data sometimes the conversion works without problem, sometime it adds a random "column28".
Is there some additional synchronizing to be done when handling many large files?
autogenerate_column_names is true by default. If the header row has 27 entries and a later row has 28 then that is what it will be called.
Which version of logstash are you running?
Thank you for your reply.
I am aware of the auto generation feature.
I am also using autodetect names
The columns match and the pipeline succeeds sometimes, hence my confusion.
I am using the latest version
Mit freundlichen Grüßen
Dr.-Ing. Christoph Weber
IT-Consultant
ASTRUM IT GmbH
Am Wolfsmantel 2
D-91058 Erlangen
Tel.: +49 91 31 94 08-309
Geschäftsführer:
Gerhard Pölz
Walter Greul
Amtsgericht Fürth, HRB 6549
UST.-Id.: DE186320885
The only thing I can think of is that pipeline.ordered is set to false, so that sometimes it picks up a data row as a header.
Unfortunately the pipeline.workers are already set to 1.
Ordering is then set by default.

© 2020. All Rights Reserved - Elasticsearch
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant logo are trademarks of the Apache Software Foundation in the United States and/or other countries.