[moe] full test for deepseek and mixtral (pp + sp to fix)

This commit is contained in:
hxwang
2024-07-19 06:11:11 +00:00
committed by Hongxin Liu
parent 7077d38d5a
commit 803878b2fd
17 changed files with 430 additions and 517 deletions

View File

@@ -9,7 +9,7 @@ import torch.nn.functional as F
from colossalai.legacy.moe.load_balance import LoadBalancer
from colossalai.legacy.moe.utils import create_ep_hierarchical_group, get_noise_generator
from colossalai.moe.operators import AllGather, AllToAll, HierarchicalAllToAll, MoeCombine, MoeDispatch, ReduceScatter
from colossalai.moe._operation import AllGather, AllToAll, HierarchicalAllToAll, MoeCombine, MoeDispatch, ReduceScatter
from colossalai.shardformer.layer.moe import MLPExperts
from colossalai.tensor.moe_tensor.api import get_dp_group, get_ep_group, get_ep_group_ranks, get_ep_size