I have a field in which I store very textual data that may reach 4000 characters, which are symbols, numbers, and (/,*,-,_,.). When I store this data in the default state, it is normal and receives the data, but when I search in term, it does not return any value. So I changed "ignore_above" to 4000 so that I can search accurately using term, but I notice that there is a delay and the use of more space.
Is there another way to store this data with a length of 4000 characters and search for it accurately using term?
I would say that is expected as you are storing a high-cardinality field with very large terms which take up a lot of space. If you are searching based on the full term, might it be an option to store a hash of the field in a separate field and search based on this?
No, I cannot divide the data because I rely on it in my search operations to a large extent, and at the same time I want the search to be as fast and accurate as possible.
No, I cannot divide the data because I rely on it in my search operations to a large extent, and at the same time I want the search to be as fast and accurate as possible.
I did not suggest dividing it but rather calculate a hash and store that in a separate field and use that for exact term lookup. This requires you to hash the query string as well and may result in hash collisions, although you may reduce the risk of that by selecting an appropriate hash function.
No, not that I am aware of.