I am using ES version 1.3.2, and Spark 1.1.0.
I can successfully read and write records from/to ES using newAPIHadoopRDD()
and saveAsNewAPIHadoopDataset().
However, I am struggling to find a way to update records. Even I specify a
'key' in ESOutputFormat it gets ignored, as documented clearly.
So my question is : Is there a way to specify document ID and custom
routing values when writing to ES using Spark? If yes, how?
Anyone has an idea?
At least if I get to know whether this is possible or not, that'll be a
great help.
Thanks.
On Wednesday, October 1, 2014 3:46:51 PM UTC+5:30, Preeti Raj - Buchhada
wrote:
I am using ES version 1.3.2, and Spark 1.1.0.
I can successfully read and write records from/to ES using newAPIHadoopRDD()
and saveAsNewAPIHadoopDataset().
However, I am struggling to find a way to update records. Even I specify a
'key' in ESOutputFormat it gets ignored, as documented clearly.
So my question is : Is there a way to specify document ID and custom
routing values when writing to ES using Spark? If yes, how?
On Mon, Oct 13, 2014 at 12:55 PM, Preeti Raj - Buchhada pbuchhada@gmail.com wrote:
Anyone has an idea?
At least if I get to know whether this is possible or not, that'll be a
great help.
Thanks.
On Wednesday, October 1, 2014 3:46:51 PM UTC+5:30, Preeti Raj - Buchhada
wrote:
I am using ES version 1.3.2, and Spark 1.1.0.
I can successfully read and write records from/to ES using
newAPIHadoopRDD() and saveAsNewAPIHadoopDataset().
However, I am struggling to find a way to update records. Even I specify a
'key' in ESOutputFormat it gets ignored, as documented clearly.
So my question is : Is there a way to specify document ID and custom
routing values when writing to ES using Spark? If yes, how?
Thanks for your reply Costin.
However, we have a need to compute a custom ID based on concatenation of
multiple fields values and then computing the hash value. So simply
specifying 'es.mapping.id' will not help in our case.
Is there any other way?
On Monday, October 13, 2014 4:08:05 PM UTC+5:30, Costin Leau wrote:
You can the mapping options [1], namely es.mapping.id to specify the
id field of your documents.
On Mon, Oct 13, 2014 at 12:55 PM, Preeti Raj - Buchhada
<pbuc...@gmail.com <javascript:>> wrote:
Anyone has an idea?
At least if I get to know whether this is possible or not, that'll be a
great help.
Thanks.
On Wednesday, October 1, 2014 3:46:51 PM UTC+5:30, Preeti Raj - Buchhada
wrote:
I am using ES version 1.3.2, and Spark 1.1.0.
I can successfully read and write records from/to ES using
newAPIHadoopRDD() and saveAsNewAPIHadoopDataset().
However, I am struggling to find a way to update records. Even I
specify a
'key' in ESOutputFormat it gets ignored, as documented clearly.
So my question is : Is there a way to specify document ID and custom
routing values when writing to ES using Spark? If yes, how?
--
You received this message because you are subscribed to the Google
Groups
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send
an
email to elasticsearc...@googlegroups.com <javascript:>.
To view this discussion on the web visit
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.