How to update http.max_content_length in elastic cloud cluster

I am using an elastic cloud cluster. getting error:

Job aborted due to stage failure: Task 1782 in stage 1.0 failed 4 times, most recent failure: Lost task 1782.3 in stage 1.0 (TID 2191,, executor 6): org.apache.spark.util.TaskCompletionListenerException: [PUT] on [data_the_284_ingest/my_data/_bulk] failed; server[...] returned [413|Request Entity Too Large:]
    at org.apache.spark.TaskContextImpl.invokeListeners(TaskContextImpl.scala:153)
    at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:131)
    at org.apache.spark.executor.Executor$
    at java.util.concurrent.ThreadPoolExecutor.runWorker(
    at java.util.concurrent.ThreadPoolExecutor$

is any way to update setting http.max_content_length ?

Changing that would probably require you to get in contact with Elastic Cloud support. I would however start by looking into why you are sending such large requests. There is a reason the limit is in place as handling very large requests can consume a lot of resources and cause instability.

Don't think it can be updated on ECE at the moment.

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.