mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-05 19:13:01 +00:00
[test] fix gemini checkpoint and gpt test (#4620)
This commit is contained in:
@@ -102,7 +102,6 @@ def check_forward_backward(model_fn, data_gen_fn, output_transform_fn, loss_fn,
|
||||
torch.cuda.empty_cache()
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="This test will hang in CI")
|
||||
@parameterize('test_config', [{
|
||||
'tp_size': 2,
|
||||
'pp_size': 2,
|
||||
@@ -220,7 +219,7 @@ def check_gpt2_3d(rank, world_size, port):
|
||||
run_gpt2_3d_test()
|
||||
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="This test will hang in CI")
|
||||
@pytest.mark.dist
|
||||
@rerun_if_address_is_in_use()
|
||||
@clear_cache_before_run()
|
||||
|
Reference in New Issue
Block a user