mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-17 23:18:36 +00:00
[colotensor] add Tensor.view op and its unit test (#1343)
[colotensor] add megatron initialization for gpt2
This commit is contained in:
@@ -1,7 +1,7 @@
|
||||
import torch
|
||||
import torch.distributed as dist
|
||||
from colossalai.tensor import ColoTensor, ColoTensorSpec
|
||||
from colossalai.tensor.distspec import _DistSpec
|
||||
from colossalai.tensor.distspec import _DistSpec, DistPlacementPattern
|
||||
|
||||
|
||||
def gather_tensor(colo_tensor: ColoTensor) -> None:
|
||||
@@ -26,7 +26,7 @@ def gather_tensor(colo_tensor: ColoTensor) -> None:
|
||||
def scatter_tensor(colo_tensor: ColoTensor, dist_spec: _DistSpec) -> None:
|
||||
"""Reversal operation of `gather_tensor`.
|
||||
"""
|
||||
if dist_spec.placement == 'r':
|
||||
if dist_spec.placement == DistPlacementPattern.REPLICATE:
|
||||
dist.broadcast(colo_tensor.data, 0)
|
||||
else:
|
||||
global_size = colo_tensor.size_global()
|
||||
|
Reference in New Issue
Block a user