org.apache.spark.SparkClassNotFoundException: [DATA_SOURCE_NOT_FOUND] Failed to find the data source: org.elasticsearch.spark.sql. Make sure the provider name is correct and the package is properly registered and compatible with your Spark version

Hi Everyone,

Currently I want to write pyspark dataframe to Elasticcloud index

Below is the code, however pyspark is unable to find the org.elasticsearch.spark.sql.

Environment - Databricks
Databricks Run time - 16.3

(   
    df
    .write
    .mode('overwrite')
    .format("org.elasticsearch.spark.sql")
    .option("es.resource", "INDEX_NAME")
    .option("es.write.operation", "index")
    .option("es.net.http.header.Authorization", "TOKEN")
    .option("es.nodes", conf['host'])
    .option("es.port", conf['port'])
    .save()
    )

Error

 org.apache.spark.SparkClassNotFoundException: [DATA_SOURCE_NOT_FOUND] Failed to find the data source: org.elasticsearch.spark.sql. Make sure the provider name is correct and the package is properly registered and compatible with your Spark version

Please provide you suggestion.