mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-03 10:06:44 +00:00
[shardformer/sequence parallel] support gpt2 seq parallel with pp/dp/tp (#4460)
* support gpt2 seq parallel with pp/dp/tp * fix a bug when waiting for stream done * delete unused gpt2_seq file
This commit is contained in:
@@ -235,6 +235,10 @@ class HybridParallelPlugin(PipelinePluginBase):
|
||||
assert dist.get_world_size() % (
|
||||
tp_size * pp_size
|
||||
) == 0, f'world size {dist.get_world_size()} is not divisible by tp_size {tp_size} * pp_size {pp_size}'
|
||||
|
||||
if enable_sequence_parallelism:
|
||||
assert tp_size > 1, 'Sequence parallelism must be enabled when using tensor parallelism'
|
||||
|
||||
# TODO(ver217): support zero
|
||||
assert zero_stage == 0, 'zero is not support yet'
|
||||
self.tp_size = tp_size
|
||||
|
Reference in New Issue
Block a user