I have recently started learning/using elasticsearch and I am trying to index some documents using NEST (c# client) but after I index one document or run a bulk index the read_only_allow_delete is set to true in the index settings. So next call for indexing documents results in 403. If I set this to false again, it is set back to true again after the next call.
Does anybody have an idea why this is happening?
Have I set up the index wrong?
I know about the solution explained here but next time I index a document it is set back to true again
Do the Elasticsearch logs indicate anything e.g. Elasticsearch enforcing a read-only index block (index.blocks.read_only_allow_delete: true) on every index that has one or more shards allocated on a node that has at least one disk exceeding the flood stage setting cluster.routing.allocation.disk.watermark.flood_stage, which defaults to 95% of disk space used?
Thanks. That was actually the problem. I re-read the document about read only allow delete flag and it'd mentioned the disk space issue. So clearing up some files did the trick .
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.