From 7af46ab6676f14726cd336eef8ea74fc9c3541bd Mon Sep 17 00:00:00 2001 From: flybird11111 <1829166702@qq.com> Date: Thu, 17 Apr 2025 17:59:46 +0800 Subject: [PATCH] fix --- .../test_kernels/cuda/test_flash_decoding_attention.py | 1 - 1 file changed, 1 deletion(-) diff --git a/tests/test_infer/test_kernels/cuda/test_flash_decoding_attention.py b/tests/test_infer/test_kernels/cuda/test_flash_decoding_attention.py index c93055fec..e9bf24d53 100644 --- a/tests/test_infer/test_kernels/cuda/test_flash_decoding_attention.py +++ b/tests/test_infer/test_kernels/cuda/test_flash_decoding_attention.py @@ -197,7 +197,6 @@ except ImportError: print("The subsequent test requires vllm. Please refer to https://github.com/vllm-project/vllm") -@clear_cache_before_run() @pytest.mark.skipif(not HAS_VLLM, reason="requires vllm") @pytest.mark.parametrize("BATCH_SIZE", [1, 7, 32]) @pytest.mark.parametrize("BLOCK_SIZE", [6, 32])