mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-05 19:13:01 +00:00
[shardformer] supported fused qkv checkpoint (#4073)
This commit is contained in:
@@ -15,7 +15,7 @@ from colossalai.nn import init as init
|
||||
from colossalai.nn.layer.utils import divide
|
||||
from colossalai.tensor.d_tensor.api import shard_colwise, shard_rowwise, sharded_tensor_to_param
|
||||
|
||||
from ._operation import gather_forward_split_backward, reduce_input
|
||||
from ._operation import gather_forward_split_backward, reduce_forward
|
||||
from .parallel_module import ParallelModule
|
||||
from .utils import create_randomizer_with_offset
|
||||
|
||||
@@ -276,5 +276,5 @@ class VocabParallelEmbedding1D(ParallelModule):
|
||||
# Mask the output embedding.
|
||||
output_parallel[input_mask, :] = 0.
|
||||
# Reduce across all the model parallel GPUs.
|
||||
output = reduce_input(output_parallel, self.process_group)
|
||||
output = reduce_forward(output_parallel, self.process_group)
|
||||
return output
|
||||
|
Reference in New Issue
Block a user