I'm looking for a way to use the (for example path_hierarchy) tokenizer(s) in a ingest pipeline. Is there anyway of doing this?
the tokenizer are only applied before a document is indexed. The ingest pipeline modifies the JSON before indexing starts (this is also the reason why you can dedicated ingest nodes, as these functions are split from each other).
I guess you could come up with a
script processor that is doing some splitting based on a character and thus create something similar than the path hierarchy tokenizer.
Thanks! Might be nice to be able to leverage the function behind the tokenizers instead of building them again in regex