Hi,
My ML configuration is 2 nodes with this configuration for each node :
ml.allocated_processors_double 8.0
ml.machine_memory 62.4GB
ml.config_version 12.0.0
ml.max_jvm_size 4GB
ml.allocated_processors 8
I already use model bge-large-en-v1.5 for our RAG but now I want to test baai_bge_m3 BGE-M3 — BGE documentation because this model can deal with more 512 tokens.
The upload of the model baai_bge_m3 is OK.
I stopped all runnings models so all resources are free and I tried to start the model baai_bge_m3. But I have this issue.
{
"error": {
"root_cause": [
{
"type": "status_exception",
"reason": "Could not start deployment because no ML nodes with sufficient capacity were found"
}
],
"type": "status_exception",
"reason": "Could not start deployment because no ML nodes with sufficient capacity were found",
"caused_by": {
"type": "illegal_state_exception",
"reason": "Could not start deployment because no suitable nodes were found, allocation explanation [none]"
}
},
"status": 429
}
I suspect “sufficient capacity” means no enough memory ???
When I use bge-large-en-v1 with 4 allocations/1 thread “Model size stats” is :
model_size_bytes 1.2GB
required_native_memory_bytes 10.2GB
BGE-M3 documentation says model size is 2,27 GB but I have no idea how much memory is needed to run this model on ELK.
I appreciate any help
Regards,