is there any other way to force custom tokenization and lemmatization besides writing a custom Token Filter plugin?
I would like to synchronize products of custom analysis (done first) with data stored in elasticsearch using a text type (done second). My aim is to use elastic for fulltext search and highlight.
I've been playing with the ingest node, but I can't see any way to force tokenization/lemmatization using processors. (I was hoping to combine somehow the two streams of tokens and lemmas into a searchable text field.) Do I have the right impression?
Thanks in advance for any suggestions or tips,