mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-24 22:42:15 +00:00
Merge 9383c84f26
into 46ed5d856b
This commit is contained in:
commit
10f0869a0d
@ -12,9 +12,7 @@ from transformers.models.llama.modeling_llama import (
|
||||
LlamaAttention,
|
||||
LlamaConfig,
|
||||
LlamaDecoderLayer,
|
||||
LlamaDynamicNTKScalingRotaryEmbedding,
|
||||
LlamaForCausalLM,
|
||||
LlamaLinearScalingRotaryEmbedding,
|
||||
LlamaMLP,
|
||||
LlamaModel,
|
||||
LlamaRMSNorm,
|
||||
|
@ -18,3 +18,4 @@ datasets
|
||||
pydantic
|
||||
ray
|
||||
peft>=0.7.1
|
||||
transformers==4.51.3
|
||||
|
@ -2,7 +2,7 @@ import pytest
|
||||
import torch
|
||||
from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM
|
||||
|
||||
from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM
|
||||
# from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM
|
||||
from colossalai.inference.spec.drafter import Drafter
|
||||
from colossalai.utils import get_current_device
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user