Please note that tokenizer
has no effect here.
Next time, could you share a script like:
POST _analyze
{
"analyzer": "french",
"text": [
"Village"
]
}
So it's easier to copy and paste and test.
the french
and italian
analyzers are coming from Lucene. I can't tell if the behavior is normal or not. But what is the problem with that behavior?