From cf50cae7e999f8836412fa24f3a11b54e0190b0b Mon Sep 17 00:00:00 2001 From: flybird11111 <1829166702@qq.com> Date: Fri, 25 Apr 2025 11:43:21 +0800 Subject: [PATCH] fix --- colossalai/inference/modeling/models/glide_llama.py | 4 ++-- tests/test_infer/test_drafter.py | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/colossalai/inference/modeling/models/glide_llama.py b/colossalai/inference/modeling/models/glide_llama.py index 0ee78a303..871dd018d 100644 --- a/colossalai/inference/modeling/models/glide_llama.py +++ b/colossalai/inference/modeling/models/glide_llama.py @@ -12,9 +12,9 @@ from transformers.models.llama.modeling_llama import ( LlamaAttention, LlamaConfig, LlamaDecoderLayer, - LlamaDynamicNTKScalingRotaryEmbedding, + # LlamaDynamicNTKScalingRotaryEmbedding, LlamaForCausalLM, - LlamaLinearScalingRotaryEmbedding, + # LlamaLinearScalingRotaryEmbedding, LlamaMLP, LlamaModel, LlamaRMSNorm, diff --git a/tests/test_infer/test_drafter.py b/tests/test_infer/test_drafter.py index 3c5dda157..bb330786b 100644 --- a/tests/test_infer/test_drafter.py +++ b/tests/test_infer/test_drafter.py @@ -2,7 +2,7 @@ import pytest import torch from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM -from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM +# from colossalai.inference.modeling.models.glide_llama import GlideLlamaConfig, GlideLlamaForCausalLM from colossalai.inference.spec.drafter import Drafter from colossalai.utils import get_current_device