This question is similar to this one:
I like to share my experience and in the same time hope I can get some
The query was run against an index with about 700 million documents.
Two things happens,
The node run this query crashed. It is the node configured not to
The data nodes start crazy on GC. eventually old generation gc cannot
reduce the heep usage and the nodes becomes unresponsive. in some cases.
OLD generation gc even increased size of the heap:
2014-12-20 07:21:03,370][WARN ][mon…
The difference between the tickets is that I have intentionally specified
size parameter. Index contains little more than 8 billions of documents. Here's my query:
My question is why did this happen?
size: 5000 only returns 5000 documents in the search response (which also is a lot and you should consider using scroll searches). However in order to get those 5000 documents the script score query needs to be executed against every hit that matches the query in the function score, which may be far more than those 5k.
This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.