I've run into an issue which is preventing me from moving forwards with ES.
I've got an application where I keep 'live' documents in ElasticSearch.
Each document is a combination from data from multiple sources, which are
merged together using doc_as_upsert. Each document has a TTL which is
updated whenever new data comes in for a document, so documents die
whenever no data source has given information about it for a while. The
amount of documents generally doesn't exceed 15.000 so it's a fairly small
Whenever I leave this running, slowly but surely memory usage on the box
creeps up, seemingly unbounded until there is no more resident memory left.
The Java process nicely keeps within its set ES_MAX_HEAP bounds, but it
seems the mapping from storage on disk to memory is every-increasing, even
when the amount of 'live' documents goes to 0.
I was wondering if anyone has seen such a memory problem before and whether
there are ways to debug memory usage which is unaccounted for by processes
You received this message because you are subscribed to the Google Groups "elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email to email@example.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/68ac8858-9074-43f1-9ad4-666de8cba344%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.