-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Closed
Description
Is it the total num of gpus or the number of gpus on a single node?
I have seen contradictory signals in the code.
accelerate/docs/source/usage_guides/ipex.md
Line 139 in ee00467
num_processes: 16 |
accelerate/src/accelerate/state.py
Line 154 in ee00467
self.num_processes = torch.distributed.get_world_size() |
here, it seems like the total number of gpus.
accelerate launch --num_processes $GPUS_PER_NODE $SCRIPT $SCRIPT_ARGS |
here, it sees like the number of gpus per node.
Metadata
Metadata
Assignees
Labels
No labels