mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-22 09:59:38 +00:00
moved env variables to global variables; (#215)
added branch context; added vocab parallel layers; moved split_batch from load_batch to tensor parallel embedding layers; updated gpt model; updated unit test cases; fixed few collective communicator bugs
This commit is contained in:
@@ -9,4 +9,4 @@ from ._sequence_parallel_gradient_handler import SequenceParallelGradientHandler
|
||||
|
||||
__all__ = ['BaseGradientHandler', 'DataParallelGradientHandler',
|
||||
'ZeROGradientHandler', 'PipelineSharedModuleGradientHandler',
|
||||
'MoeGradientHandler', 'SequenceParallelGradientHandler']
|
||||
'MoeGradientHandler', 'SequenceParallelGradientHandler']
|
@@ -9,7 +9,6 @@ from typing import Iterable, Callable
|
||||
from .._base_engine import Engine
|
||||
from colossalai.logging import get_dist_logger
|
||||
from colossalai.utils import get_current_device
|
||||
from colossalai.nn.layer import split_batch
|
||||
|
||||
|
||||
class BaseSchedule(ABC):
|
||||
@@ -69,7 +68,6 @@ class BaseSchedule(ABC):
|
||||
self.batch_size = data.size(0)
|
||||
else:
|
||||
self.batch_size = next(iter(data.values())).size(0)
|
||||
data, label = split_batch(data), split_batch(label)
|
||||
if to_gpu:
|
||||
return self._move_to_device(data), self._move_to_device(label)
|
||||
return data, label
|
||||
|
Reference in New Issue
Block a user