Hi All,
I'm trying to run a Spark job that uses the Elasticsearch client directly (instead of using elasticsearch-hadoop jar). I'm running into the NoSuchFieldError that has been seen my others. I've followed the advice on those posts to ensure that there are no other jackson jars making it on to the classpath and yet the error persists.
$ mvn dependency:tree | grep jackson (after some clean up)
com.fasterxml.jackson.core:jackson-annotations:jar:2.6.2:compile
com.fasterxml.jackson.core:jackson-core:jar:2.6.2:compile
com.fasterxml.jackson.core:jackson-databind:jar:2.6.2:compile
com.fasterxml.jackson.dataformat:jackson-dataformat-yaml:jar:2.6.2:compile
com.fasterxml.jackson.module:jackson-module-paranamer:jar:2.6.2:test
com.fasterxml.jackson.module:jackson-module-scala_2.10:jar:2.6.2:test
org.codehaus.jackson:jackson-core-asl:jar:1.8.8:provided
org.codehaus.jackson:jackson-jaxrs:jar:1.8.3:provided
org.codehaus.jackson:jackson-mapper-asl:jar:1.8.8:provided
org.codehaus.jackson:jackson-xc:jar:1.8.3:provided
org.json4s:json4s-jackson_2.10:jar:3.2.10:provided
I did notice that the older codehaus jars are also there (in the provided scope coming from hadoop and spark cdh5.8.0), but since the package name has changed, it shouldn't pose a problem.
Please advise.
Thanks,
Pradeep