I'm looking for a way to use the (for example path_hierarchy) tokenizer(s) in a ingest pipeline. Is there anyway of doing this?
the tokenizer are only applied before a document is indexed. The ingest pipeline modifies the JSON before indexing starts (this is also the reason why you can dedicated ingest nodes, as these functions are split from each other).
I guess you could come up with a
script processor that is doing some splitting based on a character and thus create something similar than the path hierarchy tokenizer.
Thanks! Might be nice to be able to leverage the function behind the tokenizers instead of building them again in regex
Please go ahead and open an issue in the elasticsearch repo about this, explaining the rationale behind it!
This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.