distributed — Model Optimizer 0.27.1 (original) (raw)
Utility functions for using torch.distributed.
Functions
backend | Returns the distributed backend. |
---|---|
barrier | Synchronizes all processes. |
get_data_parallel_group | Deprecated method. |
get_tensor_parallel_group | Deprecated method. |
is_available | Returns whether the distributed package is available. |
is_initialized | Returns whether the distributed package is initialized. |
is_master | Returns whether the current process is the master process. |
rank | Returns the rank of the current process. |
set_data_parallel_group | Deprecated method. |
set_tensor_parallel_group | Deprecated method. |
size | Returns the number of processes. |
backend()
Returns the distributed backend.
Return type:
str | None
barrier(group=None)
Synchronizes all processes.
Return type:
None
get_data_parallel_group()
Deprecated method.
Return type:
None
get_tensor_parallel_group()
Deprecated method.
Return type:
None
is_available()
Returns whether the distributed package is available.
Return type:
bool
is_initialized()
Returns whether the distributed package is initialized.
Return type:
bool
is_master(group=None)
Returns whether the current process is the master process.
Return type:
bool
rank(group=None)
Returns the rank of the current process.
Return type:
int
set_data_parallel_group(group)
Deprecated method.
set_tensor_parallel_group(group)
Deprecated method.
size(group=None)
Returns the number of processes.
Return type:
int