hi,
i have a heavy index into which i am indexing documents.. lot of times, i
am updating existing docs with new/updated field values..
when i run the stats api, i see that the index is accumulating bunch of
deletes over time even though no explicit delete api is called. how is that
possible? is there something happening under the covers?
hi,
i have a heavy index into which i am indexing documents.. lot of times, i
am updating existing docs with new/updated field values..
when i run the stats api, i see that the index is accumulating bunch of
deletes over time even though no explicit delete api is called. how is that
possible? is there something happening under the covers?
i thought so too.. but then i created a test index with a test document and
updated it.. running stats on test index didn't show any deletes. hence i
got confused.
assuming that updates causes deletes, does it make sense to have a
daily/weekly cron job to run optimize with expunge deletes option to keep
the index lightweight all the time? is that a recommended practice? if not,
what are the recommendations in this regard?
thanks
On Wed, Nov 21, 2012 at 3:24 PM, Ivan Brusic ivan@brusic.com wrote:
Updates in Lucene are essentially deletes + inserts.
hi,
i have a heavy index into which i am indexing documents.. lot of times, i
am updating existing docs with new/updated field values..
when i run the stats api, i see that the index is accumulating bunch of
deletes over time even though no explicit delete api is called. how is that
possible? is there something happening under the covers?
i thought so too.. but then i created a test index with a test document
and updated it.. running stats on test index didn't show any deletes. hence
i got confused.
assuming that updates causes deletes, does it make sense to have a
daily/weekly cron job to run optimize with expunge deletes option to keep
the index lightweight all the time? is that a recommended practice? if not,
what are the recommendations in this regard?
thanks
On Wed, Nov 21, 2012 at 3:24 PM, Ivan Brusic ivan@brusic.com wrote:
Updates in Lucene are essentially deletes + inserts.
hi,
i have a heavy index into which i am indexing documents.. lot of times,
i am updating existing docs with new/updated field values..
when i run the stats api, i see that the index is accumulating bunch of
deletes over time even though no explicit delete api is called. how is that
possible? is there something happening under the covers?
i thought so too.. but then i created a test index with a test document
and updated it.. running stats on test index didn't show any deletes. hence
i got confused.
assuming that updates causes deletes, does it make sense to have a
daily/weekly cron job to run optimize with expunge deletes option to keep
the index lightweight all the time? is that a recommended practice? if not,
what are the recommendations in this regard?
thanks
On Wed, Nov 21, 2012 at 3:24 PM, Ivan Brusic <iv...@brusic.com<javascript:>
wrote:
Updates in Lucene are essentially deletes + inserts.
hi,
i have a heavy index into which i am indexing documents.. lot of times,
i am updating existing docs with new/updated field values..
when i run the stats api, i see that the index is accumulating bunch of
deletes over time even though no explicit delete api is called. how is that
possible? is there something happening under the covers?
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.