I need some backup policies for my environment.
I am using 5 clusters with 3 nodes which is spread on 3 machines.
I need to take backup of all indexes in a cluster.
So what i thinking is shall i copy and archive the "data" directory alone..is this correct ..?
Also its enough to take the backup from any one node for a cluster, Whether using that i can able to recover the cluster ..?
After your backup is done, you should enable flush back again.
You only need to do that one one node if that node contains a complete set
of shards. For example, if you have 3 nodes, and all your indices have 2
replicas, then it should work. If you have 0 or one replicas, it's best to
backup all data directories.
Here are two scripts I've tested for backing up and restoring "old" indices
containing logs, in case it gives any insight:
I was also optimizing them, because they wouldn't change from that point
on. I don't know it it makes sense for you or not.
I need some backup policies for my environment.
I am using 5 clusters with 3 nodes which is spread on 3 machines.
I need to take backup of all indexes in a cluster.
So what i thinking is shall i copy and archive the "data" directory
alone..is this correct ..?
Also its enough to take the backup from any one node for a cluster, Whether
using that i can able to recover the cluster ..?
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.