I started a rolling restart yesterday but has add to stop because the disks
were filling up oddly. It looks like when the bode comes up it no longer
deletes shards it can't use.
Elasticsearch reports that the disk is nearly full but that it isn't using
most of the space. When I look myself the disk is mostly full and most of
the space is taken up by shards.
I'm not clear where to go from here though. Find the files elasticsearch
doesn't have open and delete them?
On Thursday, August 21, 2014 2:44:19 PM UTC+2, Nikolas Everett wrote:
I started a rolling restart yesterday but has add to stop because the
disks were filling up oddly. It looks like when the bode comes up it no
longer deletes shards it can't use.
Elasticsearch reports that the disk is nearly full but that it isn't using
most of the space. When I look myself the disk is mostly full and most of
the space is taken up by shards.
I'm not clear where to go from here though. Find the files elasticsearch
doesn't have open and delete them?
Hi Nikolas,
Can you provide the output of curl 'localhost:9200/_cat/shards?v' and curl 'localhost:9200/_cat/health?v'? Also, can you describe your cluster
topology and what the current disk usages are for all nodes across the
cluster?
Additionally, what version of ES are you using before and after the upgrade?
On Thursday, August 21, 2014 2:44:19 PM UTC+2, Nikolas Everett wrote:
I started a rolling restart yesterday but has add to stop because the
disks were filling up oddly. It looks like when the bode comes up it no
longer deletes shards it can't use.
Elasticsearch reports that the disk is nearly full but that it isn't
using most of the space. When I look myself the disk is mostly full and
most of the space is taken up by shards.
I'm not clear where to go from here though. Find the files elasticsearch
doesn't have open and delete them?
Hi Nikolas,
Can you provide the output of curl 'localhost:9200/_cat/shards?v' and curl 'localhost:9200/_cat/health?v'? Also, can you describe your cluster
topology and what the current disk usages are for all nodes across the
cluster?
Additionally, what version of ES are you using before and after the
upgrade?
Total - free on disk is 479163707392
But used is 238902736642
Meaning about 50% of used space isn't accounted for.
But everything on that partition is in elasticsearch's directory:
manybubbles@elastic1001:/var/lib/elasticsearch/production-search-eqiad/nodes/0/indices$
du -h | tail -n1
447G .
Its like when we did the upgrade some files weren't deleted when they were
no longer in use.
On Thu, Aug 21, 2014 at 10:24 AM, Nikolas Everett nik9000@gmail.com wrote:
Hi Lee! Thanks for responding. Ok, here goes:
Version: 1.2.1->1.3.2
curl 'localhost:9200/_cat/health?v:
epoch timestamp cluster status node.total node.data
shards pri relo init unassign
1408630877 14:21:17 production-search-eqiad green 17 17
6050 2017 0 0 0
On Thursday, August 21, 2014 2:44:19 PM UTC+2, Nikolas Everett wrote:
I started a rolling restart yesterday but has add to stop because the
disks were filling up oddly. It looks like when the bode comes up it no
longer deletes shards it can't use.
Elasticsearch reports that the disk is nearly full but that it isn't
using most of the space. When I look myself the disk is mostly full and
most of the space is taken up by shards.
I'm not clear where to go from here though. Find the files elasticsearch
doesn't have open and delete them?
Hi Nikolas,
Can you provide the output of curl 'localhost:9200/_cat/shards?v' and curl 'localhost:9200/_cat/health?v'? Also, can you describe your cluster
topology and what the current disk usages are for all nodes across the
cluster?
Additionally, what version of ES are you using before and after the
upgrade?
On Thu, Aug 21, 2014 at 10:35 AM, Nikolas Everett nik9000@gmail.com wrote:
This gist shows the error in action: gist:3acdb38052dba3fbc5a0 · GitHub
Total - free on disk is 479163707392
But used is 238902736642
Meaning about 50% of used space isn't accounted for.
But everything on that partition is in elasticsearch's directory:
manybubbles@elastic1001:/var/lib/elasticsearch/production-search-eqiad/nodes/0/indices$
du -h | tail -n1
447G .
Its like when we did the upgrade some files weren't deleted when they were
no longer in use.
On Thu, Aug 21, 2014 at 10:24 AM, Nikolas Everett nik9000@gmail.com
wrote:
Hi Lee! Thanks for responding. Ok, here goes:
Version: 1.2.1->1.3.2
curl 'localhost:9200/_cat/health?v:
epoch timestamp cluster status node.total node.data
shards pri relo init unassign
1408630877 14:21:17 production-search-eqiad green 17
17 6050 2017 0 0 0
On Thursday, August 21, 2014 2:44:19 PM UTC+2, Nikolas Everett wrote:
I started a rolling restart yesterday but has add to stop because the
disks were filling up oddly. It looks like when the bode comes up it no
longer deletes shards it can't use.
Elasticsearch reports that the disk is nearly full but that it isn't
using most of the space. When I look myself the disk is mostly full and
most of the space is taken up by shards.
I'm not clear where to go from here though. Find the files
elasticsearch doesn't have open and delete them?
Hi Nikolas,
Can you provide the output of curl 'localhost:9200/_cat/shards?v' and curl 'localhost:9200/_cat/health?v'? Also, can you describe your cluster
topology and what the current disk usages are for all nodes across the
cluster?
Additionally, what version of ES are you using before and after the
upgrade?
Side note: after digging through the code for two hours I can't find
anything that sweeps up files/directories/local shard storage that is
unused. I see lots of deletes done in finally blocks but I'm not sure how
I got in this state nor if there is something designed to dig me out of it.
For posterity: if you nuke the contents of your node's disk after stopping
Elasticsearch 1.2 but before starting Elasticsearch 1.3 then you won't end
up with too much data that can't be cleared. The more nodes you upgrade the
more shards you'll be able to delete any way. https://github.com/s1monw
On Thu, Aug 21, 2014 at 2:37 PM, Nikolas Everett nik9000@gmail.com wrote:
Side note: after digging through the code for two hours I can't find
anything that sweeps up files/directories/local shard storage that is
unused. I see lots of deletes done in finally blocks but I'm not sure how
I got in this state nor if there is something designed to dig me out of it.
For posterity: if you nuke the contents of your node's disk after
stopping Elasticsearch 1.2 but before starting Elasticsearch 1.3 then
you won't end up with too much data that can't be cleared. The more
nodes you upgrade the more shards you'll be able to delete any
way.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.