mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-08 12:30:42 +00:00
[zero] refactor ShardedParamV2 for convenience (#742)
This commit is contained in:
@@ -76,7 +76,7 @@ def _run_test_sharded_optim_v2(cpu_offload,
|
||||
# check whether parameters are identical in ddp
|
||||
for name, p in zero_model.named_parameters():
|
||||
if not p.colo_attr.param_is_sharded and p.colo_attr.is_replicated:
|
||||
assert_equal_in_group(p.colo_attr.sharded_data_tensor.payload.to(get_current_device()))
|
||||
assert_equal_in_group(p.colo_attr.data_payload.to(get_current_device()))
|
||||
|
||||
model = MoeModel(checkpoint=True).half()
|
||||
col_model_deepcopy(zero_model, model)
|
||||
@@ -100,7 +100,7 @@ def _run_test_sharded_optim_v2(cpu_offload,
|
||||
for (n, p), zp in zip(apex_model.named_parameters(), zero_model.parameters()):
|
||||
if 'gate' in n:
|
||||
p.data = p.float()
|
||||
p.data.copy_(zp.colo_attr.sharded_data_tensor.payload)
|
||||
p.data.copy_(zp.colo_attr.data_payload)
|
||||
|
||||
for i, (data, label) in enumerate(train_dataloader):
|
||||
if i > 5:
|
||||
|
Reference in New Issue
Block a user