mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-20 09:01:06 +00:00
[pipeline/chimera] test chimera | fix bug of initializing (#1615)
* [pipeline/tuning] improve dispatch performance both time and space cost * [pipeline/converge] add interface for testing convergence * [NFC] polish colossalai/utils/multi_tensor_apply/multi_tensor_apply.py code style * Update PipelineBase.py * [pipeline/chimera] reconstruct PipelineBase and Worker to support more feasible custom schedule | finish Chimera * [pipeline/chimera] test chimera | fix bug of initializing
This commit is contained in:
@@ -110,7 +110,8 @@ class PipelinableContext(InsertPostInitMethodToModuleSubClasses):
|
||||
name_list.append((name, param))
|
||||
|
||||
for name, param in name_list:
|
||||
delattr(module, name)
|
||||
if hasattr(module, name):
|
||||
delattr(module, name)
|
||||
setattr(module, name, ColoParameter.from_torch_tensor(tensor=param.data, requires_grad=param.requires_grad))
|
||||
|
||||
def to_layer_list(self, exec_seq=None):
|
||||
|
Reference in New Issue
Block a user