Force external tokenization/lemmatization


is there any other way to force custom tokenization and lemmatization besides writing a custom Token Filter plugin?

I would like to synchronize products of custom analysis (done first) with data stored in elasticsearch using a text type (done second). My aim is to use elastic for fulltext search and highlight.

I've been playing with the ingest node, but I can't see any way to force tokenization/lemmatization using processors. (I was hoping to combine somehow the two streams of tokens and lemmas into a searchable text field.) Do I have the right impression?

Thanks in advance for any suggestions or tips,


This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.