Hi there,
I want to be able to provide a text search against a document that contains a large number (10+ digits) but running into some conversion issues with my mapping.
Scenario: User can submit a value of 1234567890 in a field and then wants to be able to return that document by searching for the exact text "1234567890".
In my current implementation, I'm storing the value in a double typed field (since the field could contain floating point values) with a sub field that copies the string value to another field with a custom analyser that splits the number by decimal point:
"numberField": {
"type": "double",
"fields": {
"asText": {
"type": "string",
"copy_to": [
"all_numbers"
]
}
},
"all_numbers": {
"type": "string",
"analyzer": "number_analyzer"
}
So this means if the value is 31, then it's stored as 31.0 in numberField and then in all_numbers as "31" and "0". So a text search against the all_numbers field will return this document.
However, when the number value is very large, it seems the double conversion to string is using the 1.XEY notation, e.g. 123456789 becomes 1.23456789E9. This means what's stored in my all_numbers field is "1" and "23456789E9" and so a text search of "1234567890" returns 0 hits.
At this point I'm just wondering whether my source document should contain the value twice, once as a number and once as string and avoid needing to do any custom analysis, however this does mean moving this logic to the application layer when I'd rather Elastic could just handle it.
Any other suggestions on how I could get Elastic to store the value as I desire?
FYI I have other custom analysers running on the _all field that means searching that field doesn't return the 'number' either.
TIA
Craig