mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2026-05-04 01:48:43 +00:00
* [booster] added the plugin base and torch ddp plugin * polish code * polish code * polish code
52 lines
1.1 KiB
Python
52 lines
1.1 KiB
Python
from abc import ABC, abstractmethod
|
|
from typing import Callable, List, Tuple, Union
|
|
|
|
import torch.nn as nn
|
|
from torch.optim import Optimizer
|
|
from torch.optim.lr_scheduler import _LRScheduler as LRScheduler
|
|
from torch.utils.data import DataLoader
|
|
|
|
from colossalai.booster.interface import OptimizerWrapper
|
|
|
|
__all__ = ['Plugin']
|
|
|
|
|
|
class Plugin(ABC):
|
|
|
|
@property
|
|
@abstractmethod
|
|
def supported_devices(self) -> List[str]:
|
|
pass
|
|
|
|
@property
|
|
@abstractmethod
|
|
def supported_precisions(self) -> List[str]:
|
|
pass
|
|
|
|
@property
|
|
@abstractmethod
|
|
def control_precision(self) -> bool:
|
|
pass
|
|
|
|
@property
|
|
@abstractmethod
|
|
def control_device(self) -> bool:
|
|
pass
|
|
|
|
@property
|
|
@abstractmethod
|
|
def support_no_sync(self) -> bool:
|
|
pass
|
|
|
|
@abstractmethod
|
|
def configure(
|
|
self,
|
|
model: nn.Module,
|
|
optimizer: Optimizer,
|
|
criterion: Callable = None,
|
|
dataloader: DataLoader = None,
|
|
lr_scheduler: LRScheduler = None,
|
|
) -> Tuple[Union[nn.Module, OptimizerWrapper, LRScheduler, DataLoader]]:
|
|
# implement this method
|
|
pass
|