Cannot detect ES version - typically this happens when accessing a WAN/Cloud instance without the proper setting 'es.nodes.wan.only' on Docker Deployment

I am using dockers for spark and elasticsearch. I wrote the following code from tutorial:

import org.apache.spark.SparkConf
import org.elasticsearch.spark._

val conf = sc.getConf
conf.set("", "true")
conf.set("es.nodes", "")
// conf.set("es.nodes.wan.only","true")
// conf.set("es.nodes.client.only","true")
// conf.set("es.nodes.discovery","true")

val rdd = sc.esRDD("index/type", "?q=me*")
val size = rdd.collect().size

but I got the following error:

import org.apache.spark.SparkConf
import org.elasticsearch.spark._
conf: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res56: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res57: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res58: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res59: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res60: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res61: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
res62: org.apache.spark.SparkConf = org.apache.spark.SparkConf@64cac11e
rdd: org.apache.spark.rdd.RDD[(String, scala.collection.Map[String,AnyRef])] = ScalaEsRDD[3] at RDD at AbstractEsRDD.scala:17
org.elasticsearch.hadoop.EsHadoopIllegalArgumentException: Cannot detect ES version - typically this happens when accessing a WAN/Cloud instance without the proper setting 'es.nodes.wan.only'
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions$lzycompute(AbstractEsRDD.scala:61)
at org.elasticsearch.spark.rdd.AbstractEsRDD.esPartitions(AbstractEsRDD.scala:60)
at org.elasticsearch.spark.rdd.AbstractEsRDD.getPartitions(AbstractEsRDD.scala:27)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)
at scala.Option.getOrElse(Option.scala:120)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:47)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:52)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:54)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:56)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:58)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:60)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:62)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:64)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:66)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:68)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:70)
at $iwC$$iwC$$iwC$$iwC$$iwC.(:72)
at $iwC$$iwC$$iwC$$iwC.(:74)
at $iwC$$iwC$$iwC.(:76)
at $iwC$$iwC.(:78)
at $iwC.(:80)
at (:82)
at .(:86)
at .()
at .(:7)
at .()
at $print()
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(

As mentioned in the code (commented) I set several parameters based on suggestions on web, but non of them worked.

This is my docker-compose file for spark and elasticsearch containers:

  container_name: spark
  hostname: spark
  image: ###
    - "8088:8088"
    - "8042:8042"
    - "7077:7077"
    - "5666:5666"
  command: /opt/src/analytics/src/main/spark/etc/ -d
    - elasticsearch
  container_name: elasticsearch
  image: ###
  command: elasticsearch
  hostname: elasticsearch
    - "9200:9200"
    - "9300:9300"

I'll be thankful if anyone can help me.

I saw the following line in the error msg:

Caused by: Connection error (check network and/or proxy settings)- all nodes failed; tried [[]]

It tried connecting elasticsearch on but it should be (due to docker deployment), anyone knows how I can config it? I already use the following conf for SparkContext:
conf.set("es.nodes", "")

Full error msg: es-hadoop.log · GitHub is not a valid IP. Or rather it is used to indicate bind all interfaces but that's not what the JVM will do - it's best to indicate the exact IP of your network interface otherwise it is implementation dependent and something tells me it will simply pick up the first interface it finds, typically the localhost.

dear @costin thank for the reply, I tried the exact IP too but it didn't work. The problem was it doesn't use the SparkContext configs! It seems, when I set the parameters, somehow it didn't set properly. I used the following code instead and it worked.

import org.elasticsearch.spark._

var esconf = Map("es.nodes" -> "x.x.x.x", "es.port" -> "9200")
EsSpark.saveToEs(rdd, "spark/docs", esconf)
val rdd = sc.esRDD("spark/docs", "?q=*", esconf)

Looks like you are using an old version. This looks like a bug that was fixed a couple of releases ago. .make sure to use the latest version.