I'm using Logstash with elasticsearch to index data. I don't remember setting up snapshot for my indices. But today when I look at the data folder, there is a snapshot directory created.
Any idead how this can happen? I'm using elasticsearch 1.5.0 and logstash 1.5.0
@warkolm , I killed my old es cluster. Cleared up all the data. Elastic search ran without any problems for 1 day. After that I started see snapshots along with elasticsearch in data folders. I know for sure that nobody created a snapshot repo.
I have attached my .yml file config in the post. Is this expected behaviour? Tell me if you need more details.
If a snapshot repository exists in your cluster someone must have added it. Elasticsearch does not add snapshot repositories automatically. Did you change your cluster name from the default (elasticsearch)? Is your cluster open to internet access? Both of these put your cluster in a vunerable state and leave it open to unintentional or malicious attacks. See https://www.elastic.co/blog/scripting-security for more details on basic security for your cluster.
just out of curiosity, you have 50% for the index_buffer_size, have the node encounter oom before? how many es nodes do you have in the cluster? what's the use case like?
@colings86 I changed my cluster name to QuickrunCluster. Es clusters are in a secure network. I don't think they are vulnerable to attack. I deleted the snapshot created and removed the snapshot template. But it keeps on getting created. Thanks!
@warkolm I have updated the details in the post. The logs are what I got. One weird thing is everytime I delete snapshot, it gets created after an hour. Can you please help me debug this?
Also regarding you second comment, can you please elaborate that. Thanks for the help.
This is the snapshot template that keeps on getting created,
What is quickrun, chances are it's something some has setup to create this so it might pay to ask your colleagues.
I'd also suggest putting either Shield in place to log what is happening, or some other reverse proxy.
If you are using time based indices then why bother with TTL? You can just delete the index after the period it needs to exist for. TTL is inefficient as ES has to constantly scan all your documents to see if a TTL has been reached.
quickrun is the project I am working on. I am currently managing these clusters. I am absolutely sure that no one else is doing this. It keeps happening again and again. I tried deleting everything and starting it again. It doesn't work. Please help me debug this. Also I run these machines in VPN.
Please help me solve this, I am so frustrated by this issue. Thanks
I don't have shield plugin, but I am running elasticsearch and logstash in a local machine(which is not connected to internet) to see if the same behavior happens. Is there any reason this can happen?
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.