ML jobs are almost always "constrained" by the amount of memory required to run - and that's a function of how many unique time-series being modeled. In general, the conservative rule of thumb is a 100MB base overhead, plus 20k-30k of RAM for every partition field. So, a job running on 5000 unique time-series could require about 250MB of RAM.
The default "limit" of RAM per ML job is 1024mb as set by the parameter model_memory_limit (see docs)
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.