io/console not supported; tty will not be manipulated
Settings: Default pipeline workers: 4
Pipeline main started
Exception in thread "Ruby-0-Thread-11: Q:/softwares/ElasticSearch/logstash-2.3.2
/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.2-java/lib/logstash/pipeline.rb:
444" Exception in thread "Ruby-0-Thread-6: Q:/softwares/ElasticSearch/logstash-2
.3.2/vendor/bundle/jruby/1.9/gems/logstash-output-elasticsearch-2.6.2-java/lib/l
ogstash/outputs/elasticsearch/buffer.rb:78"
Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler i
n thread "Ruby-0-Thread-6: Q:/softwares/ElasticSearch/logstash-2.3.2/vendor/bund
le/jruby/1.9/gems/logstash-output-elasticsearch-2.6.2-java/lib/logstash/outputs/
elasticsearch/buffer.rb:78"
java.lang.OutOfMemoryError: GC overhead limit exceeded
Exception in thread "Ruby-0-Thread-3: Q:/softwares/ElasticSearch/logstash-2.3.2/
vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/task.rb:22" java.lang.OutOfMem
oryError: GC overhead limit exceeded
Exception in thread "[main]<jdbc" java.lang.OutOfMemoryError: GC overhead limit
exceeded
Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler i
n thread "[main]>worker0"
Exception in thread "[main]>worker0" Exception in thread "[main]>worker3" Except
ion in thread "[main]>worker2" java.lang.OutOfMemoryError: GC overhead limit exc
eeded
Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler i
n thread "[main]>worker2"
java.lang.ArrayIndexOutOfBoundsException: -1
at org.jruby.runtime.ThreadContext.popRubyClass(ThreadContext.java:702)
at org.jruby.runtime.ThreadContext.postYield(ThreadContext.java:1266)
Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler i
n thread "[main]>worker3"
Exception in thread "[main]>worker1" java.lang.OutOfMemoryError: GC overhead lim
it exceeded
Thanks @warkolm for the suggestion.
When I reduce the size to 50000, it is working.
But I have more than 50,00,000 records.
Can you please suggest any other way...
Ok I'll stream it.
In order to stream it, I'm thinking to stream it for every 5,00,000 records in 10 logstash instances at a time inorder to index 5,000,000 records.
will it reduce the indexing speed???
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.