In our use case, ( ES 7.4.0 ) we will iterate over multiple data ( text, pdf, etc ). Assume we have
(name, status, salary, designation, location) are necessary and mandatory fields we require from each data.
Apart from this each data might have their own fields/information. ( etc doc1 might have extra information like ( hobbies, address ) . doc2 might have ( hobbies, gender, no_kids ) etc. In the real use case these informations are a lot and vary.
The problem is, now we are indexing all these fields into the same index. The no of fields are getting more in number. But, our core algorithm which uses this index, is only making use of the (name, status, salary, designation, location) this information. Other extra information is not that useful as of now.
Is the current approach of allowing unlimited fields into index efficient?