mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-07 03:52:01 +00:00
[devops] fix compatibility (#5444)
* [devops] fix compatibility * [hotfix] update compatibility test on pr * [devops] fix compatibility * [devops] record duration during comp test * [test] decrease test duration * fix falcon
This commit is contained in:
@@ -182,7 +182,7 @@ class MoeHybridParallelPlugin(HybridParallelPlugin):
|
||||
overlap_communication: bool = True,
|
||||
use_ep_inside: bool = True,
|
||||
custom_policy: Policy = None,
|
||||
checkpoint_io: Optional[MoECheckpintIO] = None,
|
||||
checkpoint_io: Optional[MoECheckpointIO] = None,
|
||||
) -> None:
|
||||
assert (
|
||||
dist.get_world_size() % (tp_size * pp_size) == 0
|
||||
@@ -341,7 +341,6 @@ class MoeHybridParallelPlugin(HybridParallelPlugin):
|
||||
**_kwargs,
|
||||
)
|
||||
|
||||
|
||||
def get_checkpoint_io(self) -> MoECheckpointIO:
|
||||
if self.checkpoint_io is None:
|
||||
self.checkpoint_io = MoECheckpointIO(self.dp_group, self.pp_group, self.tp_group, self.zero_stage)
|
||||
|
Reference in New Issue
Block a user