this is my error log:

[parent] Data too large, 
data for [indices:data/write/bulk[s]] would be [30380624822/28.2gb], 
which is larger than the limit of [28975929753/26.9gb], 
real usage: [30380571432/28.2gb], 
new bytes reserved: [53390/52.1kb], 
usages [request=0/0b, fielddata=7605882/7.2mb, in_flight_requests=89910/87.8kb, model_inference=0/0b, accounting=77912706/74.3mb]

my cluster settings are:
heap.max = 29.9gb = true = 90%
so the limit is: 29.9gb x 90% =[28975929753/26.9gb]

my question is why the real usage[30380571432/28.2gb] can bigger than the limit[28975929753/26.9gb] ?
and how can i find out who uses this memory: usage[30380571432/28.2gb]

HI @casterQ

There are documentation that explains the circuit breaker errors step by step:

Fix common cluster issues | Elasticsearch Guide [7.12] | Elastic

Maybe I can help you understand

Best regards

Thanks for the documentation;
My question is that at some point,current real usage is larger than the limit(heap.max *,Shouldn't it breaker when the limit is reached?

Looking forward to your recovery

Your bulk sizes are too large to fit into heap, so look at decreasing those.

i think my bulk size only need 52.1KB mem,and currnet use 28.2gb is larger than limit 26.9gb

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.