Since the RDD is created lazily, there's no validation done during its
declaration - the actual logic behind it is executed when an operation
(like count) is performed. It looks like the nodes are not properly
accessed; potentially the port is not being picked up.
Can you try passing the es.node configuration through the SparkConf
object and see whether that makes a difference? If that doesn't work,
please enable logging on the rest package [1] and check whether the
proper nodes (and ports) are properly accessed.
If that's not the case, please raise an issue [2]
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.