mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-07-17 17:02:04 +00:00
fix
This commit is contained in:
parent
397875e640
commit
28cf1e2c57
@ -1,4 +1,5 @@
|
|||||||
import torch
|
import torch
|
||||||
|
import pytest
|
||||||
import torch.distributed as dist
|
import torch.distributed as dist
|
||||||
from torch.distributed.distributed_c10d import _get_default_group
|
from torch.distributed.distributed_c10d import _get_default_group
|
||||||
from torch.testing import assert_close
|
from torch.testing import assert_close
|
||||||
@ -36,6 +37,7 @@ def run_dist(rank, world_size, port):
|
|||||||
check_4gpu()
|
check_4gpu()
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.skip("tested in corresponding sharderformer")
|
||||||
@rerun_if_address_is_in_use()
|
@rerun_if_address_is_in_use()
|
||||||
def test_all_gather():
|
def test_all_gather():
|
||||||
spawn(run_dist, 4)
|
spawn(run_dist, 4)
|
||||||
|
Loading…
Reference in New Issue
Block a user