Upping to 1GB, memory usage seems to level off at 750MB, but there's a
problem in there somewhere. I'm getting a failure message, and the marvel
dashboard isn't able to fetch.
C:\elasticsearch-1.1.1\bin>elasticsearch
Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true
[2014-07-30 17:33:27,138][INFO ][node ] [Mondo]
version[1.1.1], pid[10864], build[f1585f0/2014-04-16
T14:27:12Z]
[2014-07-30 17:33:27,139][INFO ][node ] [Mondo]
initializing ...
[2014-07-30 17:33:27,163][INFO ][plugins ] [Mondo] loaded
[ldap-river, marvel], sites [marvel]
[2014-07-30 17:33:30,731][INFO ][node ] [Mondo]
initialized
[2014-07-30 17:33:30,731][INFO ][node ] [Mondo]
starting ...
[2014-07-30 17:33:31,027][INFO ][transport ] [Mondo]
bound_address {inet[/0.0.0.0:9300]}, publish_address
{inet[/192.168.0.6:9300]}
[2014-07-30 17:33:34,202][INFO ][cluster.service ] [Mondo]
new_master [Mondo][liyNQAHAS0-8f-qDDqa5Rg][twilson-T
HINK][inet[/192.168.0.6:9300]], reason: zen-disco-join (elected_as_master)
[2014-07-30 17:33:34,239][INFO ][discovery ] [Mondo]
elasticsearch/liyNQAHAS0-8f-qDDqa5Rg
[2014-07-30 17:33:34,600][INFO ][http ] [Mondo]
bound_address {inet[/0.0.0.0:9200]}, publish_address
{inet[/192.168.0.6:9200]}
[2014-07-30 17:33:35,799][INFO ][gateway ] [Mondo]
recovered [66] indices into cluster_state
[2014-07-30 17:33:35,815][INFO ][node ] [Mondo] started
[2014-07-30 17:33:39,823][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:39,830][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:39,837][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:39,838][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:43,973][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:44,212][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:44,357][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:44,501][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:53,294][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:53,309][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:53,310][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:33:53,310][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:34:03,281][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:34:03,283][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:34:03,286][DEBUG][action.search.type ] [Mondo] All
shards failed for phase: [query_fetch]
[2014-07-30 17:34:45,662][ERROR][marvel.agent.exporter ] [Mondo] create
failure (index:[.marvel-2014.07.31] type: [no
de_stats]): UnavailableShardsException[[.marvel-2014.07.31][0] [2] shardIt,
[0] active : Timeout waiting for [1m], reque
st: org.elasticsearch.action.bulk.BulkShardRequest@39b65640]
On Wednesday, July 30, 2014 5:30:29 PM UTC-7, Mark Walkom wrote:
Up that to 1GB and see if it starts.
512MB is pretty tiny, you're better off starting at 1/2GB if you can.
Regards,
Mark Walkom
Infrastructure Engineer
Campaign Monitor
email: ma...@campaignmonitor.com <javascript:>
web: www.campaignmonitor.com
On 31 July 2014 10:28, Tom Wilson <twils...@gmail.com <javascript:>>
wrote:
JDK 1.7.0_51
It has 512MB of heap, which was enough -- I've been running it like that
for the past few months, and I only have two indexes and around 300-400
documents. This is a development instance I'm running on my local machine.
This only happened when I started it today.
-tom
On Wednesday, July 30, 2014 5:16:11 PM UTC-7, Mark Walkom wrote:
What java version? How much heap have you allocated and how much RAM on
the server?
Basically you have too much data for the heap size, so increasing it
will help.
Regards,
Mark Walkom
Infrastructure Engineer
Campaign Monitor
email: ma...@campaignmonitor.com
web: www.campaignmonitor.com
On 31 July 2014 10:11, Tom Wilson twils...@gmail.com wrote:
Help! Elasticsearch was working fine, but now it's using up all its
heap space in the matter of a few minutes. I uninstalled the river and am
performing no queries. How do I diagnose the problem? 2-3 minutes after
starting, it runs out of heap space, and I'm not sure how to find out why.
Here is the profile of memory usage:
https://lh6.googleusercontent.com/-La0i_IrQBLA/U9mIyZZDYLI/AAAAAAAAFx0/SfnYVdKvFAw/s1600/elasticsearch-memory.png
And here is the console output. You can see it takes less than a
minute after starting to run out of memory. This isn't even enough time to
examine the indices through marvel.
C:\elasticsearch-1.1.1\bin>elasticsearch
Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true
[2014-07-30 16:59:02,579][INFO ][node ] [Texas
Twister] version[1.1.1], pid[8572], build[f1585f0/201
4-04-16T14:27:12Z]
[2014-07-30 16:59:02,580][INFO ][node ] [Texas
Twister] initializing ...
[2014-07-30 16:59:02,600][INFO ][plugins ] [Texas
Twister] loaded [marvel], sites [marvel]
[2014-07-30 16:59:06,437][INFO ][node ] [Texas
Twister] initialized
[2014-07-30 16:59:06,437][INFO ][node ] [Texas
Twister] starting ...
[2014-07-30 16:59:06,691][INFO ][transport ] [Texas
Twister] bound_address {inet[/0.0.0.0:9300]}, publish
_address {inet[/192.168.0.6:9300]}
[2014-07-30 16:59:09,862][INFO ][cluster.service ] [Texas
Twister] new_master [Texas Twister][ShQRhZRFQnuZMTRCu
vY9XQ][twilson-THINK][inet[/192.168.0.6:9300]], reason: zen-disco-join
(elected_as_master)
[2014-07-30 16:59:09,902][INFO ][discovery ] [Texas
Twister] elasticsearch/ShQRhZRFQnuZMTRCuvY9XQ
[2014-07-30 16:59:10,213][INFO ][http ] [Texas
Twister] bound_address {inet[/0.0.0.0:9200]}, publish
_address {inet[/192.168.0.6:9200]}
[2014-07-30 16:59:11,631][INFO ][gateway ] [Texas
Twister] recovered [65] indices into cluster_state
java.lang.OutOfMemoryError: Java heap space
Dumping heap to java_pid8572.hprof ...
Heap dump file created [814218130 bytes in 14.202 secs]
Exception in thread "elasticsearch[Texas Twister][generic][T#2]"
java.lang.OutOfMemoryError: Java heap space
at java.lang.Class.getDeclaredFields0(Native Method)
at java.lang.Class.privateGetDeclaredFields(Class.java:2397)
at java.lang.Class.getDeclaredFields(Class.java:1806)
at org.apache.lucene.util.RamUsageEstimator.
shallowSizeOfInstance(RamUsageEstimator.java:388)
at org.apache.lucene.codecs.lucene42.
Lucene42DocValuesProducer.(Lucene42DocValuesProducer.java:101)
at org.apache.lucene.codecs.lucene42.Lucene42NormsFormat.
normsProducer(Lucene42NormsFormat.java:75)
at org.apache.lucene.index.SegmentCoreReaders.(
SegmentCoreReaders.java:123)
at org.apache.lucene.index.SegmentReader.(
SegmentReader.java:96)
at org.apache.lucene.index.ReadersAndUpdates.getReader(
ReadersAndUpdates.java:141)
at org.apache.lucene.index.ReadersAndUpdates.getReadOnlyClone(
ReadersAndUpdates.java:235)
at org.apache.lucene.index.StandardDirectoryReader.open(
StandardDirectoryReader.java:101)
at org.apache.lucene.index.IndexWriter.getReader(
IndexWriter.java:382)
at org.apache.lucene.index.DirectoryReader.open(
DirectoryReader.java:111)
at org.apache.lucene.search.SearcherManager.(
SearcherManager.java:89)
at org.elasticsearch.index.engine.internal.InternalEngine.
buildSearchManager(InternalEngine.java:1471)
at org.elasticsearch.index.engine.internal.
InternalEngine.start(InternalEngine.java:279)
at org.elasticsearch.index.shard.service.InternalIndexShard.
performRecoveryPrepareForTranslog(InternalIndexShard
.java:699)
at org.elasticsearch.index.gateway.local.
LocalIndexShardGateway.recover(LocalIndexShardGateway.java:205)
at org.elasticsearch.index.gateway.IndexShardGatewayService$1.
run(IndexShardGatewayService.java:197)
at java.util.concurrent.ThreadPoolExecutor.runWorker(
ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(
ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
--
You received this message because you are subscribed to the Google
Groups "elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send
an email to elasticsearc...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/
msgid/elasticsearch/21e9cc63-0c5d-4ea0-96a2-78d817b89236%
40googlegroups.com
https://groups.google.com/d/msgid/elasticsearch/21e9cc63-0c5d-4ea0-96a2-78d817b89236%40googlegroups.com?utm_medium=email&utm_source=footer
.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an
email to elasticsearc...@googlegroups.com <javascript:>.
To view this discussion on the web visit
https://groups.google.com/d/msgid/elasticsearch/662fdcc9-0ed0-4547-aaf1-26f12efaef91%40googlegroups.com
https://groups.google.com/d/msgid/elasticsearch/662fdcc9-0ed0-4547-aaf1-26f12efaef91%40googlegroups.com?utm_medium=email&utm_source=footer
.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups "elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscribe@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/31aa9efb-7a5a-47b3-86f6-2ac9de61edb1%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.