Hi,
I am using default tokenizer(standard) for my index in elastic search. and adding documents to it. but standard tokenizer can't split words which having "." dot in it. For example:
POST _analyze
{
"tokenizer": "standard",
"text": "pink.jpg"
}
Gives me the response as:
{
"tokens": [
{
"token": "pink.jpg",
"start_offset": 0,
"end_offset": 8,
"type": "<ALPHANUM>",
"position": 0
}
]
}
The above response showing the whole word in one term. Can we divide it into two terms using "."(dot) operator in standard tokenizer? is any setting in standard tokenizer for this?