mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-27 15:57:16 +00:00
fix
This commit is contained in:
parent
5276288fc8
commit
cf50cae7e9
@ -12,9 +12,9 @@ from transformers.models.llama.modeling_llama import (
|
||||
LlamaAttention,
|
||||
LlamaConfig,
|
||||
LlamaDecoderLayer,
|
||||
LlamaDynamicNTKScalingRotaryEmbedding,
|
||||
# LlamaDynamicNTKScalingRotaryEmbedding,
|
||||
LlamaForCausalLM,
|
||||
LlamaLinearScalingRotaryEmbedding,
|
||||
# LlamaLinearScalingRotaryEmbedding,
|
||||
LlamaMLP,
|
||||
LlamaModel,
|
||||
LlamaRMSNorm,
|
||||
|
@ -2,7 +2,7 @@ import pytest
|
||||
import torch
|
||||
from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM
|
||||
|
||||
from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM
|
||||
# from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM
|
||||
from colossalai.inference.spec.drafter import Drafter
|
||||
from colossalai.utils import get_current_device
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user