hi, I want use standard tokenizer
to split sentence into words first, then use synonym filter
to normalize synonyms, and finally use ngram tokenizer
to break the words that output from synonym filter into characters。In simple terms:
standard tokenizer
-> split sentences
synonym filter
-> map the synonyms to a single word,
ngram tokenizer
-> break the words that output from synonym filter into characters
Is there any way to do this?