i am new to the ELK and i would like to convert data in a field called "subject" to UTF8. The data comes pushed from logstash to elasticsearch and visualized with kibana. My problem is that the data comes in an ugly encoding and i would like to have a way to convert it to UTF8. Could anyone give me a good start how i could solve my issue. Thanks in advance,
What encoding is your data in when you index it into Elasticsearch? Elasticsearch and Kibana will support UTF-8. Can you convert it prior to indexing into Elasticsearch?
Strictly speaking it is a subject line from postfix which i parse with grok patterns. I dont know a pretty way to convert it prior pushing as postfix logging cant convert it while logging. I thought that there might be a way to convert it from or with kibana (?).
As you can see, sometimes the subject is broken becaus of german characters.
I tried to convert it just for a test from the shell with a perl oneliner: cat test.txt | perl -MEncode -ne 'print encode("utf8",decode("MIME-Header",$_))' which beautifies the subject as expected.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.