[docs] change placememt_policy to placement_policy (#3829)

* fix typo colossalai/autochunk auto_parallel amp

* fix typo colossalai/auto_parallel nn utils etc.

* fix typo colossalai/auto_parallel autochunk fx/passes  etc.

* fix typo docs/

* change placememt_policy to placement_policy in docs/ and examples/
This commit is contained in:
digger yu
2023-05-24 14:51:49 +08:00
committed by GitHub
parent e90fdb1000
commit 518b31c059
7 changed files with 20 additions and 20 deletions

View File

@@ -340,12 +340,12 @@ def get_full_repo_name(model_id: str, organization: Optional[str] = None, token:
# Gemini + ZeRO DDP
def gemini_zero_dpp(model: torch.nn.Module, placememt_policy: str = "auto"):
def gemini_zero_dpp(model: torch.nn.Module, placement_policy: str = "auto"):
from colossalai.nn.parallel import GeminiDDP
model = GeminiDDP(model,
device=get_current_device(),
placement_policy=placememt_policy,
placement_policy=placement_policy,
pin_memory=True,
search_range_mb=64)
return model