I'm trying to preserve specific characters during tokenization using word_delimiter filter by defining the type_table (as described in http://www.fullscale.co/blog/2013/03/04/preserving_specific_characters_during_tokenizing_in_elasticsearch.html).
Actually my idea is to override the built-in English analyzer by including custom configured "word_delimiter" ("type_table": ["# => ALPHA", "@ => ALPHA"]) filter, but I cannot find any way to do it.
I also tried to create a custom english analyzer but still getting next problems:
- I don't actually know the default settings of the built-in english analyzer (But I really want to preserve it)
- While trying to set "tokenizer": english getting an error on creating index, saying that english tokenizer is not found.
I'm using 0.90.5 ES
Hope for your kind help!