mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-08-06 18:43:58 +00:00
[pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
This commit is contained in:
parent
509274c47e
commit
5acd2190e6
@ -1,5 +1,6 @@
|
|||||||
from typing import Any, Dict
|
|
||||||
import copy
|
import copy
|
||||||
|
from typing import Any, Dict
|
||||||
|
|
||||||
import ray
|
import ray
|
||||||
import ray.util.collective as cc
|
import ray.util.collective as cc
|
||||||
import torch
|
import torch
|
||||||
@ -31,6 +32,7 @@ def ray_broadcast_object(obj: Any, src: int = 0, device=None, group_name: str =
|
|||||||
obj = c10d._tensor_to_object(obj, size_tensor.item())
|
obj = c10d._tensor_to_object(obj, size_tensor.item())
|
||||||
return obj
|
return obj
|
||||||
|
|
||||||
|
|
||||||
def ray_broadcast_tensor_dict(
|
def ray_broadcast_tensor_dict(
|
||||||
tensor_dict: Dict[str, torch.Tensor],
|
tensor_dict: Dict[str, torch.Tensor],
|
||||||
src: int = 0,
|
src: int = 0,
|
||||||
|
@ -7,7 +7,9 @@ import ray
|
|||||||
import ray.util.collective as cc
|
import ray.util.collective as cc
|
||||||
import torch
|
import torch
|
||||||
import torch.distributed as dist
|
import torch.distributed as dist
|
||||||
|
from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict
|
||||||
from coati.distributed.profiling_utils import CustomProfiler
|
from coati.distributed.profiling_utils import CustomProfiler
|
||||||
|
from coati.distributed.utils import bind_batch, post_recv, unbind_batch
|
||||||
from tqdm import tqdm
|
from tqdm import tqdm
|
||||||
|
|
||||||
from colossalai.booster import Booster
|
from colossalai.booster import Booster
|
||||||
@ -15,9 +17,6 @@ from colossalai.booster.plugin import HybridParallelPlugin
|
|||||||
from colossalai.initialize import launch
|
from colossalai.initialize import launch
|
||||||
from colossalai.utils import get_current_device
|
from colossalai.utils import get_current_device
|
||||||
|
|
||||||
from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict
|
|
||||||
from coati.distributed.utils import bind_batch, post_recv, unbind_batch
|
|
||||||
|
|
||||||
|
|
||||||
class BaseConsumer:
|
class BaseConsumer:
|
||||||
def __init__(
|
def __init__(
|
||||||
|
@ -3,12 +3,11 @@ import time
|
|||||||
import ray
|
import ray
|
||||||
import ray.util.collective as cc
|
import ray.util.collective as cc
|
||||||
import torch
|
import torch
|
||||||
|
from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict
|
||||||
from coati.distributed.profiling_utils import CustomProfiler
|
from coati.distributed.profiling_utils import CustomProfiler
|
||||||
|
|
||||||
from colossalai.utils import get_current_device
|
from colossalai.utils import get_current_device
|
||||||
|
|
||||||
from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict
|
|
||||||
|
|
||||||
|
|
||||||
@ray.remote
|
@ray.remote
|
||||||
class Distributor:
|
class Distributor:
|
||||||
|
@ -5,9 +5,9 @@ import ray
|
|||||||
import torch
|
import torch
|
||||||
import wandb
|
import wandb
|
||||||
from coati.distributed.comm import SharedVariableActor
|
from coati.distributed.comm import SharedVariableActor
|
||||||
from coati.distributed.zero_bubble.consumer import BaseConsumer
|
|
||||||
from coati.distributed.loss import PolicyLoss
|
from coati.distributed.loss import PolicyLoss
|
||||||
from coati.distributed.utils import memory_efficient_logprob
|
from coati.distributed.utils import memory_efficient_logprob
|
||||||
|
from coati.distributed.zero_bubble.consumer import BaseConsumer
|
||||||
from coati.trainer.utils import all_reduce_mean, all_reduce_sum
|
from coati.trainer.utils import all_reduce_mean, all_reduce_sum
|
||||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||||
|
|
||||||
|
@ -11,9 +11,12 @@ import torch
|
|||||||
import tqdm
|
import tqdm
|
||||||
import wandb
|
import wandb
|
||||||
from coati.dataset.loader import RawConversationDataset, collate_fn_grpo
|
from coati.dataset.loader import RawConversationDataset, collate_fn_grpo
|
||||||
|
from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict
|
||||||
|
from coati.distributed.inference_backend import BACKEND_MAP
|
||||||
from coati.distributed.profiling_utils import CustomProfiler
|
from coati.distributed.profiling_utils import CustomProfiler
|
||||||
from coati.distributed.reward.reward_fn import boxed_math_reward_fn, code_reward_fn, math_reward_fn
|
from coati.distributed.reward.reward_fn import boxed_math_reward_fn, code_reward_fn, math_reward_fn
|
||||||
from coati.distributed.reward.verifiable_reward import VerifiableReward
|
from coati.distributed.reward.verifiable_reward import VerifiableReward
|
||||||
|
from coati.distributed.utils import pre_send, safe_append_to_jsonl_file
|
||||||
from ray.util.collective import allreduce
|
from ray.util.collective import allreduce
|
||||||
from ray.util.collective.types import ReduceOp
|
from ray.util.collective.types import ReduceOp
|
||||||
from torch.utils.data import DataLoader, DistributedSampler
|
from torch.utils.data import DataLoader, DistributedSampler
|
||||||
@ -21,10 +24,6 @@ from transformers import AutoTokenizer
|
|||||||
|
|
||||||
from colossalai.utils import get_current_device
|
from colossalai.utils import get_current_device
|
||||||
|
|
||||||
from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict
|
|
||||||
from coati.distributed.inference_backend import BACKEND_MAP
|
|
||||||
from coati.distributed.utils import pre_send, safe_append_to_jsonl_file
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from vllm import SamplingParams
|
from vllm import SamplingParams
|
||||||
except ImportError:
|
except ImportError:
|
||||||
|
Loading…
Reference in New Issue
Block a user