Documentation Index
Fetch the complete documentation index at: https://ray-preview.mintlify.app/llms.txt
Use this file to discover all available pages before exploring further.
ScalingConfig is the smallest object you’ll touch when scaling a training job. It controls how many workers run, what each worker gets, and how they’re placed.
Basic usage
Resource customization
resources_per_worker follows the same shape as Ray’s @ray.remote(...) resource spec.
Placement strategies
Workers are scheduled into a placement group. Choose how to lay it out:| Strategy | Effect |
|---|---|
PACK (default) | Pack onto as few nodes as possible. |
SPREAD | Spread across as many nodes as possible. |
STRICT_PACK | All workers on one node, or fail. |
STRICT_SPREAD | One worker per node, or fail. |
Trainer resources
A small “trainer” actor coordinates workers. Set its resources separately:Heterogeneous workers
Ray Train doesn’t yet support different resource specs per worker out of the box. For mixed workloads (e.g., one parameter server actor + N learners), build your own coordinator on top of Ray Core.Validate the config
Next steps
Run config
Storage, naming, callbacks.
Distributed PyTorch
See ScalingConfig in real training jobs.