mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-05 19:13:01 +00:00
[doc] Fix typo under colossalai and doc(#3618)
* Fixed several spelling errors under colossalai * Fix the spelling error in colossalai and docs directory * Cautious Changed the spelling error under the example folder * Update runtime_preparation_pass.py revert autograft to autograd * Update search_chunk.py utile to until * Update check_installation.py change misteach to mismatch in line 91 * Update 1D_tensor_parallel.md revert to perceptron * Update 2D_tensor_parallel.md revert to perceptron in line 73 * Update 2p5D_tensor_parallel.md revert to perceptron in line 71 * Update 3D_tensor_parallel.md revert to perceptron in line 80 * Update README.md revert to resnet in line 42 * Update reorder_graph.py revert to indice in line 7 * Update p2p.py revert to megatron in line 94 * Update initialize.py revert to torchrun in line 198 * Update routers.py change to detailed in line 63 * Update routers.py change to detailed in line 146 * Update README.md revert random number in line 402
This commit is contained in:
@@ -71,7 +71,7 @@ class CheckpointIO(ABC):
|
||||
|
||||
Args:
|
||||
model (nn.Module): model to be loaded.
|
||||
checkpoint (str): checkpoint path. This value is made compatiblity with the model checkpoints in the
|
||||
checkpoint (str): checkpoint path. This value is made compatibility with the model checkpoints in the
|
||||
mainstream model zoos such as Hugging Face and TIMM. The checkpoint path can be:
|
||||
1. a file path, e.g. 'model.pt'
|
||||
2. a path to a json file which defines the index to the sharded checkpoint
|
||||
@@ -127,7 +127,7 @@ class CheckpointIO(ABC):
|
||||
1. a file path, e.g. 'model.pt'
|
||||
2. a directory path to save the sharded checkpoint, e.g. './checkpoints/' when shard = True.
|
||||
shard (bool): whether to shard the checkpoint. Default: False. If set to True, the checkpoint will be sharded into
|
||||
multiple files. The model shards will be specificed by a `model.index.json` file. When shard = True, please ensure
|
||||
multiple files. The model shards will be specified by a `model.index.json` file. When shard = True, please ensure
|
||||
that the checkpoint path is a directory path instead of a file path.
|
||||
gather_dtensor (bool): whether to gather the distributed tensor to the first device. Default: True.
|
||||
variant (str): If specified, weights are saved in the format pytorch_model.<variant>.bin. Default: None.
|
||||
@@ -149,7 +149,7 @@ class CheckpointIO(ABC):
|
||||
|
||||
Args:
|
||||
optimizer (Optimizer): optimizer to be loaded.
|
||||
checkpoint (str): checkpoint path. This value is made compatiblity with the model checkpoints in the
|
||||
checkpoint (str): checkpoint path. This value is made compatibility with the model checkpoints in the
|
||||
"""
|
||||
index_file_exists, index_file_path = has_index_file(checkpoint)
|
||||
|
||||
@@ -180,7 +180,7 @@ class CheckpointIO(ABC):
|
||||
2. a path to a json file which defines the index to the sharded checkpoint for the optimizer
|
||||
3. a path to a folder containing a unique .index.json file for sharded checkpoint
|
||||
shard (bool): whether to shard the checkpoint. Default: False. If set to True, the checkpoint will be sharded into
|
||||
multiple files. The optimizer shards will be specificed by a `optimizer.index.json` file.
|
||||
multiple files. The optimizer shards will be specified by a `optimizer.index.json` file.
|
||||
gather_dtensor (bool): whether to gather the distributed tensor to the first device. Default: True.
|
||||
prefix (str): prefix for the optimizer checkpoint when shard = True. Default: None.
|
||||
size_per_shard (int): size per shard in MB. Default: 1024. This value is only used when shard is set to True.
|
||||
|
Reference in New Issue
Block a user