mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-14 21:51:57 +00:00
[misc] update pre-commit and run all files (#4752)
* [misc] update pre-commit * [misc] run pre-commit * [misc] remove useless configuration files * [misc] ignore cuda for clang-format
This commit is contained in:
@@ -13,7 +13,7 @@ from colossalai.legacy.tensor import ComputePattern, ComputeSpec, ShardSpec
|
||||
|
||||
def set_seed(seed):
|
||||
random.seed(seed)
|
||||
os.environ['PYTHONHASHSEED'] = str(seed)
|
||||
os.environ["PYTHONHASHSEED"] = str(seed)
|
||||
np.random.seed(seed)
|
||||
torch.manual_seed(seed)
|
||||
torch.cuda.manual_seed(seed)
|
||||
@@ -27,12 +27,12 @@ def check_equal(A, B):
|
||||
|
||||
def replace_parameter_add_grad(layer, weight=None, bias=None):
|
||||
if weight is not None:
|
||||
delattr(layer, 'weight')
|
||||
setattr(layer, 'weight', weight)
|
||||
delattr(layer, "weight")
|
||||
setattr(layer, "weight", weight)
|
||||
layer.weight.requires_grad = True
|
||||
if bias is not None:
|
||||
delattr(layer, 'bias')
|
||||
setattr(layer, 'bias', bias)
|
||||
delattr(layer, "bias")
|
||||
setattr(layer, "bias", bias)
|
||||
layer.bias.requires_grad = True
|
||||
|
||||
|
||||
@@ -47,12 +47,9 @@ def tensor_equal(t_a: torch.Tensor, t_b: torch.Tensor, rtol: float = 1e-3, atol:
|
||||
return True
|
||||
|
||||
|
||||
def tensor_shard_equal(tensor: torch.Tensor,
|
||||
shard: torch.Tensor,
|
||||
rank: int,
|
||||
world_size: int,
|
||||
rtol: float = 1e-3,
|
||||
atol: float = 1e-1):
|
||||
def tensor_shard_equal(
|
||||
tensor: torch.Tensor, shard: torch.Tensor, rank: int, world_size: int, rtol: float = 1e-3, atol: float = 1e-1
|
||||
):
|
||||
assert tensor.ndim == shard.ndim
|
||||
if tensor.shape == shard.shape:
|
||||
return tensor_equal(tensor, shard, rtol, atol)
|
||||
|
@@ -48,17 +48,17 @@ def check_mem():
|
||||
|
||||
|
||||
def run_dist(rank, world_size, port):
|
||||
colossalai.legacy.launch(config={}, rank=rank, world_size=world_size, host='localhost', port=port, backend='nccl')
|
||||
colossalai.legacy.launch(config={}, rank=rank, world_size=world_size, host="localhost", port=port, backend="nccl")
|
||||
check_mem()
|
||||
run()
|
||||
|
||||
|
||||
@pytest.mark.dist
|
||||
@pytest.mark.parametrize('world_size', [1, 4])
|
||||
@pytest.mark.parametrize("world_size", [1, 4])
|
||||
@rerun_if_address_is_in_use()
|
||||
def test_dist_spec_mgr(world_size):
|
||||
spawn(run_dist, world_size)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
if __name__ == "__main__":
|
||||
test_dist_spec_mgr(4)
|
||||
|
@@ -9,26 +9,27 @@ from colossalai.testing import free_port
|
||||
|
||||
@pytest.mark.skip
|
||||
def test_multiinheritance():
|
||||
colossalai.legacy.launch(config={}, rank=0, world_size=1, host='localhost', port=free_port(), backend='nccl')
|
||||
colossalai.legacy.launch(config={}, rank=0, world_size=1, host="localhost", port=free_port(), backend="nccl")
|
||||
colo_param = ColoParameter(None, requires_grad=True)
|
||||
assert colo_param.dist_spec.placement.value == 'r'
|
||||
assert colo_param.dist_spec.placement.value == "r"
|
||||
assert isinstance(colo_param, ColoTensor)
|
||||
assert isinstance(colo_param, torch.nn.Parameter)
|
||||
|
||||
# __deepcopy__ overload
|
||||
import copy
|
||||
|
||||
colo_param2 = copy.deepcopy(colo_param)
|
||||
assert isinstance(colo_param2, ColoParameter)
|
||||
assert tensor_equal(colo_param.data, colo_param2.data)
|
||||
assert colo_param.requires_grad == colo_param2.requires_grad
|
||||
|
||||
# __repr__ overload
|
||||
assert 'ColoParameter' in str(colo_param)
|
||||
assert "ColoParameter" in str(colo_param)
|
||||
|
||||
# __torch_function__
|
||||
clone_param = torch.clone(colo_param)
|
||||
assert isinstance(clone_param, ColoTensor)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
if __name__ == "__main__":
|
||||
test_multiinheritance()
|
||||
|
Reference in New Issue
Block a user