mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-06 19:40:28 +00:00
Merge pull request #6071 from wangbluo/ring_attention
[Ring Attention] fix the 2d ring attn when using multiple machine
This commit is contained in:
@@ -1177,7 +1177,10 @@ class HybridParallelPlugin(PipelinePluginBase):
|
||||
gradient_checkpoint_config=gradient_checkpoint_config,
|
||||
fp8_communication=fp8_communication,
|
||||
inner_ring_size=inner_ring_size,
|
||||
pg_mesh=self.pg_mesh,
|
||||
sp_axis=self.sp_axis,
|
||||
)
|
||||
|
||||
self.amp_config = dict(
|
||||
initial_scale=initial_scale,
|
||||
growth_factor=growth_factor,
|
||||
|
Reference in New Issue
Block a user