mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-01 17:17:05 +00:00
[feat] refactored extension module (#5298)
* [feat] refactored extension module * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish
This commit is contained in:
36
extensions/__init__.py
Normal file
36
extensions/__init__.py
Normal file
@@ -0,0 +1,36 @@
|
||||
from .cpu_adam import CpuAdamArmExtension, CpuAdamX86Extension
|
||||
from .flash_attention import (
|
||||
FlashAttentionDaoCudaExtension,
|
||||
FlashAttentionNpuExtension,
|
||||
FlashAttentionXformersCudaExtension,
|
||||
)
|
||||
from .layernorm import LayerNormCudaExtension
|
||||
from .moe import MoeCudaExtension
|
||||
from .optimizer import FusedOptimizerCudaExtension
|
||||
from .softmax import ScaledMaskedSoftmaxCudaExtension, ScaledUpperTriangleMaskedSoftmaxCudaExtension
|
||||
|
||||
ALL_EXTENSIONS = [
|
||||
CpuAdamArmExtension,
|
||||
CpuAdamX86Extension,
|
||||
LayerNormCudaExtension,
|
||||
MoeCudaExtension,
|
||||
FusedOptimizerCudaExtension,
|
||||
ScaledMaskedSoftmaxCudaExtension,
|
||||
ScaledUpperTriangleMaskedSoftmaxCudaExtension,
|
||||
FlashAttentionDaoCudaExtension,
|
||||
FlashAttentionXformersCudaExtension,
|
||||
FlashAttentionNpuExtension,
|
||||
]
|
||||
|
||||
__all__ = [
|
||||
"CpuAdamArmExtension",
|
||||
"CpuAdamX86Extension",
|
||||
"LayerNormCudaExtension",
|
||||
"MoeCudaExtension",
|
||||
"FusedOptimizerCudaExtension",
|
||||
"ScaledMaskedSoftmaxCudaExtension",
|
||||
"ScaledUpperTriangleMaskedSoftmaxCudaExtension",
|
||||
"FlashAttentionDaoCudaExtension",
|
||||
"FlashAttentionXformersCudaExtension",
|
||||
"FlashAttentionNpuExtension",
|
||||
]
|
Reference in New Issue
Block a user