I am posting the logs into multiple messages as it hit the limit of message size:
And the Running Query:
parquet at NativeMethodAccessorImpl.java:0+details
org.apache.spark.sql.DataFrameWriter.parquet(DataFrameWriter.scala:557) java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base/java.lang.reflect.Method.invoke(Method.java:564) py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) py4j.Gateway.invoke(Gateway.java:282) py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) py4j.commands.CallCommand.execute(CallCommand.java:79) py4j.GatewayConnection.run(GatewayConnection.java:238) java.base/java.lang.Thread.run(Thread.java:844)
Details of Query 0:
> == Parsed Logical Plan ==
> InsertIntoHadoopFsRelationCommand hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520, false, Parquet, Map(compression -> gzip, path -> hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520), ErrorIfExists, [hxAgentLastAudit, ,apSysSipStatsActiveSubscriptions, ... 2620 more fields]
> +- Repartition 16, false
> +- Relation[hxAgentLastAudit#0,lastStep#1, ,apSysSipStatsActiveSubscriptions#98L,... 2620 more fields] ElasticsearchRelation(Map(es.nodes -> localhost:9200, es.nodes.resolve.hostname -> false, es.resource -> event-2019.03.24),org.apache.spark.sql.SQLContext@4ff5e0e8,None)
>
> == Analyzed Logical Plan ==
> InsertIntoHadoopFsRelationCommand hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520, false, Parquet, Map(compression -> gzip, path -> hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520), ErrorIfExists, [hxAgentLastAudit, ,apSysSipStatsActiveSubscriptions, ... 2620 more fields]
> +- Repartition 16, false
> +- Relation[hxAgentLastAudit#0,lastStep#1, ,apSysSipStatsActiveSubscriptions#98L,... 2620 more fields] ElasticsearchRelation(Map(es.nodes -> localhost:9200, es.nodes.resolve.hostname -> false, es.resource -> event-2019.03.24),org.apache.spark.sql.SQLContext@4ff5e0e8,None)
>
> == Optimized Logical Plan ==
> InsertIntoHadoopFsRelationCommand hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520, false, Parquet, Map(compression -> gzip, path -> hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520), ErrorIfExists, [hxAgentLastAudit, , apSysSipStatsActiveSubscriptions, ... 2620 more fields]
> +- Repartition 16, false
> +- Relation[hxAgentLastAudit#0,lastStep#1, ,apSysSipStatsActiveSubscriptions#98L,... 2620 more fields] ElasticsearchRelation(Map(es.nodes -> localhost:9200, es.nodes.resolve.hostname -> false, es.resource -> event-2019.03.24),org.apache.spark.sql.SQLContext@4ff5e0e8,None)
>
> == Physical Plan ==
> Execute InsertIntoHadoopFsRelationCommand InsertIntoHadoopFsRelationCommand hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520, false, Parquet, Map(compression -> gzip, path -> hdfs://node-master:9000/events/1/2019/03/24.parquet.0408_1520), ErrorIfExists, [hxAgentLastAudit, , apSysSipStatsActiveSubscriptions, ... 2620 more fields]
> +- Coalesce 16
> +- Scan ElasticsearchRelation(Map(es.nodes -> localhost:9200, es.nodes.resolve.hostname -> false, es.resource -> event-2019.03.24),org.apache.spark.sql.SQLContext@4ff5e0e8,None) [hxAgentLastAudit#0,lastStep#1, ,apSysSipStatsActiveSubscriptions#98L,... 2620 more fields] PushedFilters: [], ReadSchema: struct<hxAgentLastAudit:timestamp,lastStep:string,*:string,ARTClientBytes:bigint,ARTClientPackets...