mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2026-01-29 21:49:54 +00:00
[doc] Fix typo under colossalai and doc(#3618)
* Fixed several spelling errors under colossalai * Fix the spelling error in colossalai and docs directory * Cautious Changed the spelling error under the example folder * Update runtime_preparation_pass.py revert autograft to autograd * Update search_chunk.py utile to until * Update check_installation.py change misteach to mismatch in line 91 * Update 1D_tensor_parallel.md revert to perceptron * Update 2D_tensor_parallel.md revert to perceptron in line 73 * Update 2p5D_tensor_parallel.md revert to perceptron in line 71 * Update 3D_tensor_parallel.md revert to perceptron in line 80 * Update README.md revert to resnet in line 42 * Update reorder_graph.py revert to indice in line 7 * Update p2p.py revert to megatron in line 94 * Update initialize.py revert to torchrun in line 198 * Update routers.py change to detailed in line 63 * Update routers.py change to detailed in line 146 * Update README.md revert random number in line 402
This commit is contained in:
@@ -64,7 +64,7 @@ class MoeContext(metaclass=SingletonMeta):
|
||||
from colossalai.core import global_context as gpc
|
||||
self.max_ep_size = gpc.config.get('max_ep_size', self.world_size)
|
||||
assert self.world_size % self.max_ep_size == 0, \
|
||||
"Maximum epxert parallel size must be a factor of the number of GPUs"
|
||||
"Maximum expert parallel size must be a factor of the number of GPUs"
|
||||
self.min_dp_size = self.world_size // self.max_ep_size
|
||||
|
||||
# Enabling kernel optimization may raise error in some cases
|
||||
|
||||
@@ -44,7 +44,7 @@ class ParallelContext(metaclass=SingletonMeta):
|
||||
# load config from file
|
||||
self._config = None
|
||||
|
||||
# default 3D parallel args, will be overwritten during process group intialization
|
||||
# default 3D parallel args, will be overwritten during process group initialization
|
||||
self.world_size = 1
|
||||
self.data_parallel_size = 1
|
||||
self.pipeline_parallel_size = 1
|
||||
@@ -264,7 +264,7 @@ class ParallelContext(metaclass=SingletonMeta):
|
||||
"""Adds world size for `parallel_mode`.
|
||||
|
||||
Args:
|
||||
parallel_mode (:class:`colossalai.context.ParallelMode`): The parallel mode correponding to the process group
|
||||
parallel_mode (:class:`colossalai.context.ParallelMode`): The parallel mode corresponding to the process group
|
||||
world_size (int): The world size to be added
|
||||
|
||||
Raises:
|
||||
|
||||
@@ -59,23 +59,23 @@ class SeedManager:
|
||||
self._current_mode = parallel_mode
|
||||
torch.cuda.set_rng_state(self._seed_states[parallel_mode])
|
||||
|
||||
def add_seed(self, parallel_mode: ParallelMode, seed: int, overwrtie: bool = False):
|
||||
def add_seed(self, parallel_mode: ParallelMode, seed: int, overwrite: bool = False):
|
||||
"""Adds a seed to the seed manager for `parallel_mode`.
|
||||
|
||||
Args:
|
||||
parallel_mode (:class:`colossalai.context.ParallelMode`): The chosen parallel mode.
|
||||
seed (int): The seed to be added.
|
||||
overwrtie (bool, optional): Whether allows to overwrite the seed that has been set already
|
||||
overwrite (bool, optional): Whether allows to overwrite the seed that has been set already
|
||||
|
||||
Raises:
|
||||
AssertionError: Raises an AssertionError if `parallel_mode` is not an instance of :class:`colossalai.context.ParallelMode`
|
||||
or the seed for `parallel_mode` has been added.
|
||||
"""
|
||||
assert isinstance(parallel_mode, ParallelMode), 'A valid ParallelMode must be provided'
|
||||
if overwrtie is False:
|
||||
if overwrite is False:
|
||||
assert parallel_mode not in self._seed_states, f'The seed for {parallel_mode} has been added'
|
||||
elif parallel_mode in self._seed_states:
|
||||
print(f"Warnning: {parallel_mode} seed has been overwritten.", flush=True)
|
||||
print(f"Warning: {parallel_mode} seed has been overwritten.", flush=True)
|
||||
|
||||
current_state = torch.cuda.get_rng_state()
|
||||
torch.cuda.manual_seed(seed)
|
||||
|
||||
Reference in New Issue
Block a user