[Inference/SpecDec] Add Basic Drafter Model Container (#5405)

* [Infer/Fix] Fix Dependency in test - RMSNorm kernel (#5399)

fix dependency in pytest

* add drafter model container (basic ver)
This commit is contained in:
Yuanheng Zhao
2024-02-28 13:48:17 +08:00
committed by Yuanheng
parent d63c469f45
commit 5a9b05f7b2
4 changed files with 216 additions and 0 deletions

View File

@@ -0,0 +1,41 @@
import pytest
import torch
from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM
from colossalai.inference.spec.drafter import Drafter
from colossalai.utils import get_current_device
NUM_LAYERS = 2
@pytest.mark.parametrize("spec_num", [5])
def test_drafter(spec_num: int):
torch.manual_seed(123)
device = get_current_device()
toy_config = LlamaConfig(num_hidden_layers=NUM_LAYERS)
toy_config.pad_token_id = toy_config.eos_token_id
drafter_model = LlamaForCausalLM(toy_config)
drafter_model = drafter_model.eval().cuda()
tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/llama-tokenizer")
drafter = Drafter(drafter_model, tokenizer, spec_num, device=device)
input_ids = torch.randint(low=5, high=1000, size=(1, 6)).to(device)
out = drafter.speculate(input_ids, spec_num)
past_kv_length = input_ids.size(1) + spec_num - 1
assert out.speculated_length == spec_num
assert out.next_tokens.shape == (spec_num,)
assert out.logits.shape == (spec_num, len(tokenizer))
assert drafter._past_key_values[0][0].size(2) == out.past_key_values[0][0].size(2) == past_kv_length
reject_num = 3
assert reject_num <= spec_num
drafter.trim_kv_cache(reject_num)
assert drafter._past_key_values[0][0].size(2) == past_kv_length - reject_num
if __name__ == "__main__":
test_drafter(spec_num=5)