mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-17 23:18:36 +00:00
[pipeline] refactor the pipeline module (#1087)
* [pipeline] refactor the pipeline module * polish code
This commit is contained in:
@@ -1,4 +1,3 @@
|
||||
from .lambda_wrapper import LambdaWrapper
|
||||
from .pipeline_wrapper import PipelineSharedModuleWrapper
|
||||
|
||||
__all__ = ['LambdaWrapper', 'PipelineSharedModuleWrapper']
|
||||
__all__ = ['PipelineSharedModuleWrapper']
|
||||
|
@@ -1,36 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
# -*- encoding: utf-8 -*-
|
||||
|
||||
import torch.nn as nn
|
||||
|
||||
from colossalai.builder import build_layer
|
||||
from colossalai.registry import LAYERS
|
||||
|
||||
|
||||
@LAYERS.register_module
|
||||
class LambdaWrapper(nn.Module):
|
||||
"""Wrap a function to nn.Module, which takes a config of layers and can fully access them.
|
||||
|
||||
Args:
|
||||
func (``Callable``): User customed function.
|
||||
layers_cfg (dict, optional): Config of layers, defaults to None.
|
||||
"""
|
||||
|
||||
def __init__(self, func, layers_cfg: dict = None):
|
||||
super().__init__()
|
||||
self.func = func
|
||||
self.layers = self._build_layers(layers_cfg)
|
||||
|
||||
def _build_layers(self, layers_cfg: dict):
|
||||
if layers_cfg is None:
|
||||
return None
|
||||
else:
|
||||
layers = []
|
||||
|
||||
for cfg in layers_cfg:
|
||||
layer = build_layer(cfg)
|
||||
layers.append(layer)
|
||||
return layers
|
||||
|
||||
def forward(self, *args, **kwargs):
|
||||
return self.func(self, *args, **kwargs)
|
Reference in New Issue
Block a user