hi, I want use
standard tokenizer to split sentence into words first, then use
synonym filter to normalize synonyms, and finally use
ngram tokenizer to break the words that output from synonym filter into characters。In simple terms：
standard tokenizer -> split sentences
synonym filter -> map the synonyms to a single word,
ngram tokenizer -> break the words that output from synonym filter into characters
Is there any way to do this?