Please note that tokenizer has no effect here.
Next time, could you share a script like:
POST _analyze
{
"analyzer": "french",
"text": [
"Village"
]
}
So it's easier to copy and paste and test.
the french and italian analyzers are coming from Lucene. I can't tell if the behavior is normal or not. But what is the problem with that behavior?