When trying to run term_vectors query for certain document I am getting "max_bytes_length_exceeded_exception:" error. The log file is spitting following information, it is not helping with what is causing that error. Wondering if anyone faced similar error. The same query works for other documents in same index.
My Mapping related to term_vector is :
"text": {
"type": "text",
"term_vector": "with_positions_offsets_payloads",
"store" : true,
"analyzer" : "fulltext_analyzer"} } } "analysis": { "filter": { "english_stop": { "type": "stop", "ignore_case": true, "stopwords": "_english_" }, "my_word_delimiter" :{ "type": "word_delimiter", "preserve_original" : true, "split_on_numerics" : true } }, "analyzer": { "fulltext_analyzer": { "type": "custom", "tokenizer": "whitespace", "filter": [ "lowercase", "english_stop", "my_word_delimiter", "type_as_payload" ] } } }
my term_vector query is
POST index_name/type/id/_termvectors
{
"fields" : ["*"],
"offsets" : false,
"payloads" : false,
"positions" : false,
"term_statistics" : true,
"field_statistics" : false,
"filter" : {
"max_num_terms" : 10,
"min_term_freq" : 1,
"min_doc_freq" : 1,
"min_word_length": 4,
"max_word_length" : 256}
}
error message from log file:
Caused by: org.elasticsearch.common.io.stream.NotSerializableExceptionWrapper: max_bytes_length_exceeded_exception: bytes can be at most 32766 in length; got 78233
at org.apache.lucene.util.BytesRefHash.add(BytesRefHash.java:263) ~[lucene-core-6.6.1.jar:6.6.1 9aa465a89b64ff2dabe7b4d50c472de32c298683 - varunthacker - 2017-08-29 21:54:39]
at org.apache.lucene.index.memory.MemoryIndex.storeTerms(MemoryIndex.java:633) ~[lucene-memory-6.6.1.jar:6.6.1 9aa465a89b64ff2dabe7b4d50c472de32c298683 - varunthacker - 2017-08-29 21:55:32]
at org.apache.lucene.index.memory.MemoryIndex.addField(MemoryIndex.java:277) ~[lucene-memory-6.6.1.jar:6.6.1 9aa465a89b64ff2dabe7b4d50c472de32c298683 - varunthacker - 2017-08-29 21:55:32]