I have a csv with more than 300kr of rows per hr. I use filebeat to ship data into elasticsearch.
My problem is that, the sending of data is very slow like 2k of rows per 3-5mins only and sometimes it stops for a while.
Are there any config hacks which will make the data shipping faster?
Well, I guess the next step is to see if the delay is in harvesting or publishing. Have you checked the logs for filebeat and logstash? Do you have other beats sending OK?
What is the specification of your Elasticsearch cluster? What kind of hardware and storage are you using?
If you want to test if Elasticsearch is limiting throughput, you can e.g. temporarily replace the Elasticsearch output with a file output and see if that changes the throughput of data collected.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.