diff --git a/colossalai/inference/modeling/models/glide_llama.py b/colossalai/inference/modeling/models/glide_llama.py index 0ee78a303..871dd018d 100644 --- a/colossalai/inference/modeling/models/glide_llama.py +++ b/colossalai/inference/modeling/models/glide_llama.py @@ -12,9 +12,9 @@ from transformers.models.llama.modeling_llama import ( LlamaAttention, LlamaConfig, LlamaDecoderLayer, - LlamaDynamicNTKScalingRotaryEmbedding, + # LlamaDynamicNTKScalingRotaryEmbedding, LlamaForCausalLM, - LlamaLinearScalingRotaryEmbedding, + # LlamaLinearScalingRotaryEmbedding, LlamaMLP, LlamaModel, LlamaRMSNorm, diff --git a/tests/test_infer/test_drafter.py b/tests/test_infer/test_drafter.py index 3c5dda157..bb330786b 100644 --- a/tests/test_infer/test_drafter.py +++ b/tests/test_infer/test_drafter.py @@ -2,7 +2,7 @@ import pytest import torch from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM -from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM +# from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM from colossalai.inference.spec.drafter import Drafter from colossalai.utils import get_current_device