We use logstash to synchronize MySQL data to ES and use the ES client api to batch delete expired data every day at 1am,online data is maintained at around 500 Mbp,it has been running normally for more than a year.
GET /
GET /_cat/nodes?v
GET /_cat/health?v
GET /_cat/indices?v
# Optionally
GET /_cat/shards?v
If some outputs are too big, please share them on gist.github.com and link them here.
Please format your code, logs or configuration files using </> icon as explained in this guide and not the citation button. It will make your post more readable.
Or use markdown style like:
```
CODE
```
This is the icon to use if you are not using markdown format:
There's a live preview panel for exactly this reasons.
Lots of people read these forums, and many of them will simply skip over a post that is difficult to read, because it's just too large an investment of their time to try and follow a wall of badly formatted text.
GET /_cat/nodes?v
host ip heap.percent ram.percent load node.role master name
*** *** 3 34 0.01 d * node_master
GET /_cat/health?v
epoch timestamp cluster status node.total node.data shards pri relo init unassign pending_tasks max_task_wait_time active_shards_percent
1559776603 23:16:43 es yellow 1 1 5 5 0 0 5 0 - 50.0%
GET /_cat/indices?v
health status index pri rep docs.count docs.deleted store.size pri.store.size
yellow open item 5 1 105216 144169 395mb 395mb
GET /_cat/shards?v
index shard prirep state docs store ip node
item 1 p STARTED 21030 80.5mb *** node_master
item 1 r UNASSIGNED
item 3 p STARTED 21021 79.8mb *** node_master
item 3 r UNASSIGNED
item 2 p STARTED 20914 71.9mb *** node_master
item 2 r UNASSIGNED
item 4 p STARTED 21073 80.9mb *** node_master
item 4 r UNASSIGNED
item 0 p STARTED 21178 81.8mb *** node_master
item 0 r UNASSIGNED
I had similar problem with a 1.x version the problem was because of too much deleted document I had 50% more deleted document. But in your case you have ~150% deleted documents.
health status index pri rep docs.count docs.deleted
yellow open item 5 1 105216 144169
I moved my data to a daily index and the problem was solved.
We use logstash to synchronize MySQL data to ES and use the ES client api to batch delete expired data every day at 1am
You'll better to go with a daily index and delete the full old index instead of deleting documents.
I can't find back the exact post about how the deletion work but I think I read it from this blog. http://blog.mikemccandless.com/
I'm surprised that 4gb of heap with few documents you are having issues.
But this is may be caused by a bug or the GC running for a long time.
That said you are using 2.4.3 which is really old. What about upgrading to 2.4.6 at least? Better to go to 7.1.1 BTW if you can.
Yes, upgrading to the latest version is the best result, but the low version to the high version must consider all aspects of compatibility, it may take more time, currently I am using the ES in the test environment, it is like a Bomb, I don’t know when it will crash again. I hope that the online ES can recover first. I have restarted and rebuilt the index, but it still doesn't work.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.