What would happen in case of errors when indexing a bulk of jsons?
is that info put in the queue again (not ACK'ed)?
you lost all the items you have ack'ed
Although every River is a singleton, does it make sense to configure
2 rivers pointing to the same queue in order to process it in parallel?
yes to increase parallelism you can activate several rivers on the same queue: rabbitmq grant unique delivery of a message.
hit: tune the bulk_size
You mean that if the river takes 100 items to index and after having
indexed 50 the server fails, the other 50 will be lost? I'm not sure
how the Bulk Index process works, but if it is atomic, I guess that,
in case of failure, none of the 100 items will be ack'ed and they will
keep in the queue until the next try right?
You mean that if the river takes 100 items to index and after having
indexed 50 the server fails, the other 50 will be lost? I'm not sure
how the Bulk Index process works, but if it is atomic, I guess that,
in case of failure, none of the 100 items will be ack'ed and they will
keep in the queue until the next try right?
You mean that if the river takes 100 items to index and after having
indexed 50 the server fails, the other 50 will be lost? I'm not sure
how the Bulk Index process works, but if it is atomic, I guess that,
in case of failure, none of the 100 items will be ack'ed and they will
keep in the queue until the next try right?
when the river riceive a message, it acks it and then It do the bulk. acked messages if the river dies on bulk you may lost these messagges.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.