mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-22 13:41:43 +00:00
Update bert.py
This commit is contained in:
parent
17654cb6cb
commit
611c1247ba
@ -1039,7 +1039,6 @@ def get_jit_fused_bert_output_forward():
|
|||||||
|
|
||||||
# Fix the tgt_len size in sequence parallel attention:
|
# Fix the tgt_len size in sequence parallel attention:
|
||||||
# same with the one in BertSdpaSelfAttention forward in v4.51.3 transformers except the
|
# same with the one in BertSdpaSelfAttention forward in v4.51.3 transformers except the
|
||||||
# _, _, tgt_len, _ = query_layer.shape
|
|
||||||
def get_bert_sequence_parallel_attention_forward(shard_config: ShardConfig):
|
def get_bert_sequence_parallel_attention_forward(shard_config: ShardConfig):
|
||||||
from transformers.models.bert.modeling_bert import BertSdpaSelfAttention
|
from transformers.models.bert.modeling_bert import BertSdpaSelfAttention
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user