Memory status when processing

(Daniel) #1


I have one node dedicated for the ML tasks. When trying to run a big job, after some time the memory status turns to soft_limit, and after some more hard_limit. The ml node has enought resources, and its ram is under 40% when this hard limit is triggered.

What could i be missing?

(Dimitris Athanasiou) #2

Hi Daniel,

A machine learning job has a default memory limit of 4GB. You can read more about this and how to change it here.

So, in your case, as your job needs more memory and the machine you're running it on can serve this, you can simply increase the memory limit for that job.

Kind regards,

(Daniel) #3

Thanks a lot!


(system) #4

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.