Leverage Ray for distributed and parallel model building with Plexe.
max_iterations
> 1) or when individual training runs are computationally intensive.
distributed=True
flag when initializing the plexe.Model
.
distributed=True
is set:
RayExecutor
for running the generated training code; otherwise, it falls back to the default ProcessExecutor
.RayExecutor
submits the training code execution as a remote task (@ray.remote
) to the Ray cluster (or local Ray instance).max_iterations
in model.build()
is greater than 1, and the Ray cluster has sufficient resources (CPUs/GPUs), Ray can potentially execute multiple training iterations in parallel, significantly speeding up the exploration of different model candidates.distributed=True
is the primary switch, advanced Ray configuration (like cluster address, resource limits) can be managed through Ray’s standard initialization methods (ray.init(...)
) or configuration files before initializing the plexe.Model
. Plexe itself reads some Ray configurations from its internal config (plexe.config.config.ray
) which might be relevant in specific deployment scenarios, but direct ray.init()
is the most common way to configure the connection.