My elasticsearch .yml has http.max_content_length set to 1900mb. I found
this post that says Netty has a 2gb limit ( https://github.com/elasticsearch/elasticsearch/issues/2237 ) so I was
expecting to be able to index documents of approximately 2gb. Does
elasticsearch impose a limit of 1gb even though I specified >1gb? If not,
I can file an issue on github.
1gb is a very large document and it is unusual to index such sizes.
There is a limit check against the heap. In order to be able to process
such length, you need a large heap alone to store the document source.
Depending on analyzer, heap demand increases even more.
You can index documents of arbitrary length if you preprocess them first
and split them into smaller chunks.
My elasticsearch .yml has http.max_content_length set to 1900mb. I found
this post that says Netty has a 2gb limit ( Document maximum size for bulk indexing over HTTP · Issue #2237 · elastic/elasticsearch · GitHub ) so I was
expecting to be able to index documents of approximately 2gb. Does
elasticsearch impose a limit of 1gb even though I specified >1gb? If not,
I can file an issue on github.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.