[zero] refactor ShardedParamV2 for convenience (#742)

This commit is contained in:
HELSON
2022-04-13 14:54:26 +08:00
committed by GitHub
parent 340e59f968
commit 22c4b88d56
16 changed files with 98 additions and 61 deletions

View File

@@ -76,7 +76,7 @@ def _run_test_sharded_optim_v2(cpu_offload,
# check whether parameters are identical in ddp
for name, p in zero_model.named_parameters():
if not p.colo_attr.param_is_sharded and p.colo_attr.is_replicated:
assert_equal_in_group(p.colo_attr.sharded_data_tensor.payload.to(get_current_device()))
assert_equal_in_group(p.colo_attr.data_payload.to(get_current_device()))
model = MoeModel(checkpoint=True).half()
col_model_deepcopy(zero_model, model)
@@ -100,7 +100,7 @@ def _run_test_sharded_optim_v2(cpu_offload,
for (n, p), zp in zip(apex_model.named_parameters(), zero_model.parameters()):
if 'gate' in n:
p.data = p.float()
p.data.copy_(zp.colo_attr.sharded_data_tensor.payload)
p.data.copy_(zp.colo_attr.data_payload)
for i, (data, label) in enumerate(train_dataloader):
if i > 5: