I have set up elasticsearch on one node and am using the Twitter river to
index tweets. It has been going fine with almost 50M tweets indexed so far
in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon as I
switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
Hi Yogesh. I am not 100% sure of this, so if someone else posts a reply
that differs from mine you should probably go with theirs, but I think this
is correct behavior.
Unless there is another process demanding memory from the OS, there is no
harm whatsoever in having the JVM consume all available RAM. It allows
Elasticsearch to put more things in memory, serve more cache hits, etc. You
definitely do not want the JVM to constantly run garbage collection so that
it can return memory to the OS that it's just going to want to request
again as soon as the system gets busy.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter river to
index tweets. It has been going fine with almost 50M tweets indexed so far
in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon as I
switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
Are you saying JVM is using 99% of the system memory or 99% of the heap?
If it's 99% of the available heap that's bad and you will have cluster
instability. I suggest increasing your JVM heap size if you can, I can't
find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter river to
index tweets. It has been going fine with almost 50M tweets indexed so far
in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon as I
switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
Thanks Joel and mjdude. What I mean is that ES is using 99% of the heap
memory (I think since Marvel showed memory as 1 GB which corresponds to the
heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has appeared
and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the Marvel
has stopped showing me my data (it still shows that disk memory is lower so
that means the data is still on disk) and upon searching Sense shows
"IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the heap?
If it's 99% of the available heap that's bad and you will have cluster
instability. I suggest increasing your JVM heap size if you can, I can't
find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter river to
index tweets. It has been going fine with almost 50M tweets indexed so far
in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon as I
switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
When it restarted did it attach to the wrong data directory? Take a look
at _nodes/_local/stats?pretty and check the 'data' directory location. Has
the cluster recovered after the restart? Check _cluster/health?pretty as
well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the heap
memory (I think since Marvel showed memory as 1 GB which corresponds to the
heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has appeared
and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Why is this happening?!?!
On Mon, Mar 23, 2015 at 9:25 PM, <mjd...@gmail.com <javascript:>> wrote:
Are you saying JVM is using 99% of the system memory or 99% of the heap?
If it's 99% of the available heap that's bad and you will have cluster
instability. I suggest increasing your JVM heap size if you can, I can't
find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter river
to index tweets. It has been going fine with almost 50M tweets indexed so
far in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon as
I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
Thanks a lot mjdude! It does seem like it attached to the wrong data
directory.
In elasticsearch/data/tool/nodes there are two 0 and 1. My data is in 0 but
node stats shows the data directory as elasticsearch/data/tool/nodes/1.
Now, how do I change this?
When it restarted did it attach to the wrong data directory? Take a look
at _nodes/_local/stats?pretty and check the 'data' directory location. Has
the cluster recovered after the restart? Check _cluster/health?pretty as
well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the heap
memory (I think since Marvel showed memory as 1 GB which corresponds to the
heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has appeared
and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the
heap? If it's 99% of the available heap that's bad and you will have
cluster instability. I suggest increasing your JVM heap size if you can, I
can't find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter river
to index tweets. It has been going fine with almost 50M tweets indexed so
far in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon as
I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
Sometimes that happens when the new node starts up via monit or other
automated thing before the old node is fully shutdown. I'd suggest
shutting down the node and verify it's done via ps before allowing the new
node to start up. In the case of monit if the check is hitting the http
port then it'll think it's down before it actually fully quits.
On Tuesday, March 24, 2015 at 1:56:54 PM UTC-4, Yogesh wrote:
Thanks a lot mjdude! It does seem like it attached to the wrong data
directory.
In elasticsearch/data/tool/nodes there are two 0 and 1. My data is in 0
but node stats shows the data directory as elasticsearch/data/tool/nodes/1.
Now, how do I change this?
On Tue, Mar 24, 2015 at 11:02 PM, <mjd...@gmail.com <javascript:>> wrote:
When it restarted did it attach to the wrong data directory? Take a look
at _nodes/_local/stats?pretty and check the 'data' directory location. Has
the cluster recovered after the restart? Check _cluster/health?pretty as
well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the heap
memory (I think since Marvel showed memory as 1 GB which corresponds to the
heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has appeared
and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the
heap? If it's 99% of the available heap that's bad and you will have
cluster instability. I suggest increasing your JVM heap size if you can, I
can't find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter river
to index tweets. It has been going fine with almost 50M tweets indexed so
far in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon
as I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
Thanks. Mine is a one node cluster so I am simply using the XCURL to shut
it down and the doing "bin/elasticsearch -d" to start it up.
To check if it has shutdown, I try to hit it using http. So, now how do I
start it with the node/0 data directory?
There is nothing there in node/1 data directory but I don't suppose
deleting it would be the solution? (Sorry for the basic questions I am new
to this!)
Sometimes that happens when the new node starts up via monit or other
automated thing before the old node is fully shutdown. I'd suggest
shutting down the node and verify it's done via ps before allowing the new
node to start up. In the case of monit if the check is hitting the http
port then it'll think it's down before it actually fully quits.
On Tuesday, March 24, 2015 at 1:56:54 PM UTC-4, Yogesh wrote:
Thanks a lot mjdude! It does seem like it attached to the wrong data
directory.
In elasticsearch/data/tool/nodes there are two 0 and 1. My data is in 0
but node stats shows the data directory as elasticsearch/data/tool/nodes/
1.
Now, how do I change this?
When it restarted did it attach to the wrong data directory? Take a
look at _nodes/_local/stats?pretty and check the 'data' directory
location. Has the cluster recovered after the restart? Check
_cluster/health?pretty as well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the heap
memory (I think since Marvel showed memory as 1 GB which corresponds to the
heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has
appeared and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the
heap? If it's 99% of the available heap that's bad and you will have
cluster instability. I suggest increasing your JVM heap size if you can, I
can't find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter
river to index tweets. It has been going fine with almost 50M tweets
indexed so far in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon
as I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is the
default 990.75MB)
As long as it's the first ES instance starting on that node it'll grab 0
instead of 1. I don't know if you can explicitly set the data node
directory in the config.
On Tuesday, March 24, 2015 at 2:16:57 PM UTC-4, Yogesh wrote:
Thanks. Mine is a one node cluster so I am simply using the XCURL to shut
it down and the doing "bin/elasticsearch -d" to start it up.
To check if it has shutdown, I try to hit it using http. So, now how do I
start it with the node/0 data directory?
There is nothing there in node/1 data directory but I don't suppose
deleting it would be the solution? (Sorry for the basic questions I am new
to this!)
On Tue, Mar 24, 2015 at 11:32 PM, <mjd...@gmail.com <javascript:>> wrote:
Sometimes that happens when the new node starts up via monit or other
automated thing before the old node is fully shutdown. I'd suggest
shutting down the node and verify it's done via ps before allowing the new
node to start up. In the case of monit if the check is hitting the http
port then it'll think it's down before it actually fully quits.
On Tuesday, March 24, 2015 at 1:56:54 PM UTC-4, Yogesh wrote:
Thanks a lot mjdude! It does seem like it attached to the wrong data
directory.
In elasticsearch/data/tool/nodes there are two 0 and 1. My data is in 0
but node stats shows the data directory as elasticsearch/data/tool/nodes/
1.
Now, how do I change this?
When it restarted did it attach to the wrong data directory? Take a
look at _nodes/_local/stats?pretty and check the 'data' directory
location. Has the cluster recovered after the restart? Check
_cluster/health?pretty as well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the
heap memory (I think since Marvel showed memory as 1 GB which corresponds
to the heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has
appeared and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the
heap? If it's 99% of the available heap that's bad and you will have
cluster instability. I suggest increasing your JVM heap size if you can, I
can't find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter
river to index tweets. It has been going fine with almost 50M tweets
indexed so far in 13 days.
When I started indexing, the JVM usage (observed via Marvel) hovered
between 10-20%, then started remaining around 30-40% but for the past 3-4
days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as soon
as I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is
the default 990.75MB)
Thanks. Will try starting again.
By the way, what will be the impact if I just delete the node/1 (then the
node.max_local_storage_nodes to 1 and start elasticsearch) ?
As long as it's the first ES instance starting on that node it'll grab 0
instead of 1. I don't know if you can explicitly set the data node
directory in the config.
On Tuesday, March 24, 2015 at 2:16:57 PM UTC-4, Yogesh wrote:
Thanks. Mine is a one node cluster so I am simply using the XCURL to shut
it down and the doing "bin/elasticsearch -d" to start it up.
To check if it has shutdown, I try to hit it using http. So, now how do I
start it with the node/0 data directory?
There is nothing there in node/1 data directory but I don't suppose
deleting it would be the solution? (Sorry for the basic questions I am new
to this!)
Sometimes that happens when the new node starts up via monit or other
automated thing before the old node is fully shutdown. I'd suggest
shutting down the node and verify it's done via ps before allowing the new
node to start up. In the case of monit if the check is hitting the http
port then it'll think it's down before it actually fully quits.
On Tuesday, March 24, 2015 at 1:56:54 PM UTC-4, Yogesh wrote:
Thanks a lot mjdude! It does seem like it attached to the wrong data
directory.
In elasticsearch/data/tool/nodes there are two 0 and 1. My data is in 0
but node stats shows the data directory as elasticsearch/data/tool/nodes/
1.
Now, how do I change this?
When it restarted did it attach to the wrong data directory? Take a
look at _nodes/_local/stats?pretty and check the 'data' directory
location. Has the cluster recovered after the restart? Check
_cluster/health?pretty as well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the
heap memory (I think since Marvel showed memory as 1 GB which corresponds
to the heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has
appeared and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the
heap? If it's 99% of the available heap that's bad and you will have
cluster instability. I suggest increasing your JVM heap size if you can, I
can't find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter
river to index tweets. It has been going fine with almost 50M tweets
indexed so far in 13 days.
When I started indexing, the JVM usage (observed via Marvel)
hovered between 10-20%, then started remaining around 30-40% but for the
past 3-4 days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as
soon as I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is
the default 990.75MB)
Thanks guys. The previous ES had not stopped fully, so killed it, then
deleted the node/1 directory, then set the node.max_local_storage_nodes to
1 in yml.
Now everything works smoothly!
I've also increased the heap space. Though I hope ES doesn't consume the
entire 10g heap space!
On Wednesday, March 25, 2015 at 1:41:16 AM UTC+5:30, Yogesh wrote:
Thanks. Will try starting again.
By the way, what will be the impact if I just delete the node/1 (then the
node.max_local_storage_nodes to 1 and start elasticsearch) ?
As long as it's the first ES instance starting on that node it'll grab 0
instead of 1. I don't know if you can explicitly set the data node
directory in the config.
On Tuesday, March 24, 2015 at 2:16:57 PM UTC-4, Yogesh wrote:
Thanks. Mine is a one node cluster so I am simply using the XCURL to
shut it down and the doing "bin/elasticsearch -d" to start it up.
To check if it has shutdown, I try to hit it using http. So, now how do
I start it with the node/0 data directory?
There is nothing there in node/1 data directory but I don't suppose
deleting it would be the solution? (Sorry for the basic questions I am new
to this!)
Sometimes that happens when the new node starts up via monit or other
automated thing before the old node is fully shutdown. I'd suggest
shutting down the node and verify it's done via ps before allowing the new
node to start up. In the case of monit if the check is hitting the http
port then it'll think it's down before it actually fully quits.
On Tuesday, March 24, 2015 at 1:56:54 PM UTC-4, Yogesh wrote:
Thanks a lot mjdude! It does seem like it attached to the wrong data
directory.
In elasticsearch/data/tool/nodes there are two 0 and 1. My data is in
0 but node stats shows the data directory as elasticsearch/data/tool/nodes/
1.
Now, how do I change this?
When it restarted did it attach to the wrong data directory? Take a
look at _nodes/_local/stats?pretty and check the 'data' directory
location. Has the cluster recovered after the restart? Check
_cluster/health?pretty as well.
On Tuesday, March 24, 2015 at 1:01:52 PM UTC-4, Yogesh wrote:
Thanks Joel and mjdude. What I mean is that ES is using 99% of the
heap memory (I think since Marvel showed memory as 1 GB which corresponds
to the heap, my RAM is 50GB)
I've increased the ES_HEAP_SIZE to 10g. But another problem has
appeared and I'm freaked out because of it!
So, after restarting my ES (curl shutdown) to increase the heap, the
Marvel has stopped showing me my data (it still shows that disk memory is
lower so that means the data is still on disk) and upon searching Sense
shows "IndexMissingException[[my_new_twitter_river] missing]"
Are you saying JVM is using 99% of the system memory or 99% of the
heap? If it's 99% of the available heap that's bad and you will have
cluster instability. I suggest increasing your JVM heap size if you can, I
can't find it right now but I remember a blog post that used twitter as a
benchmark and they also could get to ~50M documents with the default 1G
heap.
On Sunday, March 22, 2015 at 3:30:57 AM UTC-4, Yogesh wrote:
Hi,
I have set up elasticsearch on one node and am using the Twitter
river to index tweets. It has been going fine with almost 50M tweets
indexed so far in 13 days.
When I started indexing, the JVM usage (observed via Marvel)
hovered between 10-20%, then started remaining around 30-40% but for the
past 3-4 days it has continuously been above 90%, reaching 99% at times!
I restarted elasticsearch thinking it might get resolved but as
soon as I switched it back on, the JVM usage went back to 90%.
Why is this happening and how can I remedy it? (The JVM memory is
the default 990.75MB)
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.