I used delete_by_query API to delete multiple documents and after that _forcemerge API to remove deleted documents.
However, when I use _forcemerge API it finishes instantly and disk usage is the same. Why my API call doesn't do anything and how can I debug reasons for such behavior? I tried to use debug/trace log levels but not sure which logger to look at and when I enable trace on root logger there is too much logs to find anything useful.
I know I have multiple deleted documents so I tried to use "only_expunge_deletes" and nothing changed. I also tried to force single segment per shard with "max_num_segments" but also, nothing changed.
Any suggestions?
Cluster configuration:
40 data nodes
3 master nodes
ES version 5.6.4
Daily indices (~200GB index size)
shards: 20 primary and 1 replica
There is ~100GB free per node. Index size is ~200GB.
There shouldn't be an issue with free disk, should it? I know that ES v2 executed _forcemerge without checking if there is enough disk
The documentation does refer to request parameters, so I suspect it might be expected. It would probably be worthwhile adding an example to the docs though.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.