mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-19 04:02:17 +00:00
fix beam_width
This commit is contained in:
parent
b2eb9cd186
commit
3ad1f3b78b
@ -176,8 +176,12 @@ def llama_attn_forward(
|
||||
|
||||
|
||||
def generate_padding_position_id(input_ids: torch.Tensor) -> torch.Tensor:
|
||||
# Replace this code and use a more flexible method to obtain padding_id, avoiding directly setting padding_id like this.
|
||||
padding_id = 2
|
||||
attention_mask = input_ids.ne(padding_id).long()
|
||||
position_ids = attention_mask.long().cumsum(-1) - 1
|
||||
position_ids.masked_fill_(attention_mask == 0, 1)
|
||||
return position_ids
|
||||
|
||||
# def unpad_inputs(input_ids: torch.Tensor):
|
||||
|
||||
|
@ -42,9 +42,8 @@ def beam_search_sample(
|
||||
|
||||
# NOTE: this beam search sample function is wrong now.
|
||||
"""
|
||||
|
||||
# beam_width = generation_config.best_of
|
||||
beam_width = 1
|
||||
|
||||
beam_width = generation_config.num_beams
|
||||
results = []
|
||||
if is_prompt:
|
||||
# Prompt phase.
|
||||
|
Loading…
Reference in New Issue
Block a user