For large volume of logs we are thinking of using HDFS as data repository for elastic search. apache flume is one of option suggested by few blogs on internet. Would like to get more information from community on this to come on conclusion. How do I instruct Elastic search to store data in HDFS and query/index same every time.
LOGSTASH ---> ELASTIC SEARCH ( want to use Kibana for visualization ) -- > HDFS
Thanks @magnusbaeck. ES-hadoop as far as I understood from documentation is to do search on Hadoop echo system. Where as my requirement is to just use high storage capacity of Hadoop.
I will be collecting huge logs from 'n' number of Micro services and need to store this in faster storage for better retrieval. any thought on that ?.
I am getting tilt towards mongoDb a bit, doing research on elastic and mongodb for the moment
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.