I want to load a big document that contain many jsons into an existing index in elasticsearch with pyspark.
I have a dataframe and I try to load many jsons with this command:
dataframe.write.format("org.elasticsearch.spark.sql").option("es.nodes" ,ESnode).option("es.port" ,ESport).option("es.resource" ,ESresource)
The process not failed but not load any document.
How I can load a big document that contain many jsons into ES with pyspark?