We are having one ELK setup and logs from various sources like apache, nginx, system messages etc are processed. We keep the ES data for 1 month and dump the rest to s3 bucket. For analysing the history data we need to again restore the content of s3 to elasticsearch. Now I want to integrate ELK with hadoop so that we can see the old data in kibana. I heard of elasticsearch-hadoop. Can anybody plz help how to dump all data to hadoop via elasticsearch and vice versa?
I saw on web that we can use hive/pig to get the data from hadoop to elasticsearch but what are the ways to do other way round. Can we use elasticsearch-hadoop connectors for this purpose? Can somebody give me some example/link?
Have you looked at the project reference documentation ? It explains the architecture, configuration and how to both read and write data through the connector for each library with code examples.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.