This commit is contained in:
wangbluo
2024-10-14 18:06:24 +08:00
parent 23199e34cc
commit 3201377e94
2 changed files with 3 additions and 4 deletions

View File

@@ -858,7 +858,6 @@ def get_gpt2_flash_attention_forward(shard_config: Optional[ShardConfig] = None)
sp_mode = shard_config.sequence_parallelism_mode
sp_group = shard_config.sequence_parallel_process_group
if sp_mode == "ring_attn":
attn_output = RingAttention.attention(
query,