Is there a easy way to export a some part of data(based on some filters)
from a index say for e.g from a master index to a new index. Apparently it
looks like I will have to use bulk API to query the data from the master
index (using some filters) and then I will have insert those documents
into the new index. Is there any better and easier way.
Is there a easy way to export a some part of data(based on some filters)
from a index say for e.g from a master index to a new index. Apparently it
looks like I will have to use bulk API to query the data from the master
index (using some filters) and then I will have insert those documents into
the new index. Is there any better and easier way.
So assuming I have only on shard and i am fetching total of 100,000
documents in steps of 10,000 documents then there will be 10 get calls and
this call will happen serially one after the other. Is there a mechanism by
any chance to execute this get calls in parallel.
On Wednesday, October 15, 2014 11:53:38 PM UTC+5:30, Alexandre Rafalovitch
wrote:
On 15 October 2014 13:14, Vijay Tiwary <vijaykr...@gmail.com <javascript:>>
wrote:
Is there a easy way to export a some part of data(based on some filters)
from a index say for e.g from a master index to a new index. Apparently
it
looks like I will have to use bulk API to query the data from the master
index (using some filters) and then I will have insert those documents
into
the new index. Is there any better and easier way.
--
You received this message because you are subscribed to the Google
Groups
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send
an
email to elasticsearc...@googlegroups.com <javascript:>.
To view this discussion on the web visit
Scan/scroll, when performed over multiple nodes/shards, is inherently
executed in parallel.
Scan/scroll depends on the scroll ID chain, so you have to execute a
scan/scroll sequence serially. For a single shard you can add filters to
the query in order to partition the search hits, these queries can be
executed with several scan/scroll requests in parallel from your client.
So assuming I have only on shard and i am fetching total of 100,000
documents in steps of 10,000 documents then there will be 10 get calls and
this call will happen serially one after the other. Is there a mechanism by
any chance to execute this get calls in parallel.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.