mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-21 13:11:27 +00:00
[shardformer] polish chatglm code
This commit is contained in:
parent
8620009dd7
commit
1a29e8fc29
@ -116,6 +116,9 @@ _POLICY_LIST = {
|
|||||||
# Sam
|
# Sam
|
||||||
"transformers.models.sam.modeling_sam.SamModel":
|
"transformers.models.sam.modeling_sam.SamModel":
|
||||||
PolicyLocation(file_name="sam", class_name="SamModelPolicy"),
|
PolicyLocation(file_name="sam", class_name="SamModelPolicy"),
|
||||||
|
# ChatGLM
|
||||||
|
"tests.kit.model_zoo.transformers.chatglm2_6b.modeling_chatglm.ChatGLMModel":
|
||||||
|
PolicyLocation(file_name="chatglm", class_name="ChatGLMModelPolicy"),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
from typing import Dict, Union
|
from typing import Dict, Union
|
||||||
|
|
||||||
import torch.nn as nn
|
import torch.nn as nn
|
||||||
from ....tests.kit.model_zoo.transformers.chatglm2_6b.modeling_chatglm import ChatGLMModel, GLMBlock
|
|
||||||
|
|
||||||
import colossalai.shardformer.layer as col_nn
|
import colossalai.shardformer.layer as col_nn
|
||||||
|
|
||||||
@ -9,49 +8,6 @@ from .basepolicy import ModulePolicyDescription, Policy, SubModuleReplacementDes
|
|||||||
|
|
||||||
__all__ = ['ChatGLMModelPolicy', 'ChatGLMForConditionalGenerationPolicy']
|
__all__ = ['ChatGLMModelPolicy', 'ChatGLMForConditionalGenerationPolicy']
|
||||||
|
|
||||||
class ChatGLMModelPolicy(Policy):
|
|
||||||
|
|
||||||
def config_sanity_check(self):
|
|
||||||
pass
|
|
||||||
|
|
||||||
def preprocess(self):
|
|
||||||
# Resize embedding
|
|
||||||
vocab_size = self.model.config.vocab_size
|
|
||||||
world_size = self.shard_config.tensor_parallel_size
|
|
||||||
|
|
||||||
if vocab_size % world_size != 0:
|
|
||||||
new_vocab_size = vocab_size + world_size - vocab_size % world_size
|
|
||||||
self.model.resize_token_embeddings(new_vocab_size)
|
|
||||||
|
|
||||||
return self.model
|
|
||||||
|
|
||||||
def module_policy(self) -> Dict[Union[str, nn.Module], ModulePolicyDescription]:
|
|
||||||
from ....tests.kit.model_zoo.transformers.chatglm2_6b.modeling_chatglm import ChatGLMModel, GLMBlock
|
|
||||||
|
|
||||||
policy = {}
|
|
||||||
|
|
||||||
if self.shard_config.enable_tensor_parallelism:
|
|
||||||
|
|
||||||
policy[GLMBlock] = ModulePolicyDescription(
|
|
||||||
attribute_replacement = {},
|
|
||||||
sub_module_replacement = [
|
|
||||||
# SubModuleReplacementDescription(
|
|
||||||
# suffix = "self_attention.query_key_value",
|
|
||||||
# target_module = col_nn.Linear1D_Col,
|
|
||||||
# ),
|
|
||||||
# SubModuleReplacementDescription(
|
|
||||||
# suffix = "self_attention.dense",
|
|
||||||
# target_module = col_nn.Linear1D_Row,
|
|
||||||
# )
|
|
||||||
# SubModuleReplacementDescription(
|
|
||||||
# suffix = "self_attention.core_attention.attention_dropout",
|
|
||||||
# target_module = col_nn.DropoutForParallelInput,
|
|
||||||
# )
|
|
||||||
],)
|
|
||||||
|
|
||||||
|
|
||||||
def postprocess(self):
|
|
||||||
return self.model
|
|
||||||
|
|
||||||
class ChatGLMModelPolicy(Policy):
|
class ChatGLMModelPolicy(Policy):
|
||||||
|
|
||||||
|
@ -19,6 +19,7 @@ from colossalai.testing import (
|
|||||||
from tests.kit.model_zoo import model_zoo
|
from tests.kit.model_zoo import model_zoo
|
||||||
from tests.test_shardformer.test_model._utils import build_model, run_forward
|
from tests.test_shardformer.test_model._utils import build_model, run_forward
|
||||||
|
|
||||||
|
|
||||||
def check_forward_backward(org_model, sharded_model, data_gen_fn, output_transform_fn, loss_fn):
|
def check_forward_backward(org_model, sharded_model, data_gen_fn, output_transform_fn, loss_fn):
|
||||||
# check forward
|
# check forward
|
||||||
org_output, org_loss, shard_output, shard_loss = run_forward(org_model, sharded_model, data_gen_fn,
|
org_output, org_loss, shard_output, shard_loss = run_forward(org_model, sharded_model, data_gen_fn,
|
||||||
|
Loading…
Reference in New Issue
Block a user