From 0e900ac5cdcf863a2e1f08ac9883e44f27eff5e5 Mon Sep 17 00:00:00 2001 From: flybird11111 <1829166702@qq.com> Date: Wed, 9 Apr 2025 17:29:08 +0800 Subject: [PATCH] fix --- .github/workflows/build_on_pr.yml | 2 +- tests/test_device/test_init_logical_pg.py | 2 -- tests/test_fp8/test_fp8_allgather.py | 2 +- 3 files changed, 2 insertions(+), 4 deletions(-) diff --git a/.github/workflows/build_on_pr.yml b/.github/workflows/build_on_pr.yml index ed66c04d0..e84240fa5 100644 --- a/.github/workflows/build_on_pr.yml +++ b/.github/workflows/build_on_pr.yml @@ -161,7 +161,7 @@ jobs: --ignore tests/test_infer_ops \ --ignore tests/test_legacy \ --ignore tests/test_smoothquant \ - tests/test_fp8/ + tests/ env: LD_LIBRARY_PATH: /github/home/.tensornvme/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64 LLAMA_PATH: /data/scratch/llama-tiny diff --git a/tests/test_device/test_init_logical_pg.py b/tests/test_device/test_init_logical_pg.py index a73f0af16..4be99b17c 100644 --- a/tests/test_device/test_init_logical_pg.py +++ b/tests/test_device/test_init_logical_pg.py @@ -1,4 +1,3 @@ -import pytest import torch import torch.distributed as dist from torch.distributed import ReduceOp @@ -27,7 +26,6 @@ def check_layer(rank, world_size, port): assert tensor.equal(tensor_to_check) -@pytest.mark.skip("tested in corresponding sharderformer") @pytest.mark.dist @rerun_if_address_is_in_use() def test_logical_pg(): diff --git a/tests/test_fp8/test_fp8_allgather.py b/tests/test_fp8/test_fp8_allgather.py index e6b618560..f29512182 100644 --- a/tests/test_fp8/test_fp8_allgather.py +++ b/tests/test_fp8/test_fp8_allgather.py @@ -13,7 +13,7 @@ from colossalai.testing import parameterize, rerun_if_address_is_in_use, spawn, "shape", [(3, 7, 16)], ) - +@clear_cache_before_run() @parameterize("dtype", [torch.bfloat16, torch.float16]) @parameterize("fp8_format", ["e4m3", "e5m2"]) @parameterize("async_op", [True, False])