mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2026-04-11 14:43:10 +00:00
* [shardformer] implement policy for all GPT-J models and test * [shardformer] support interleaved pipeline parallel for bert finetune * [shardformer] shardformer support falcon (#4883) * [shardformer]: fix interleaved pipeline for bert model (#5048) * [hotfix]: disable seq parallel for gptj and falcon, and polish code (#5093) * Add Mistral support for Shardformer (#5103) * [shardformer] add tests to mistral (#5105) --------- Co-authored-by: Pengtai Xu <henryxu880@gmail.com> Co-authored-by: ppt0011 <143150326+ppt0011@users.noreply.github.com> Co-authored-by: flybird11111 <1829166702@qq.com> Co-authored-by: eric8607242 <e0928021388@gmail.com>
20 lines
408 B
Python
20 lines
408 B
Python
from .albert import *
|
|
from .bert import *
|
|
from .blip2 import *
|
|
from .bloom import *
|
|
from .chatglm2 import *
|
|
from .falcon import *
|
|
from .gpt import *
|
|
from .gptj import *
|
|
from .llama import *
|
|
from .opt import *
|
|
from .sam import *
|
|
from .t5 import *
|
|
from .vit import *
|
|
from .whisper import *
|
|
|
|
try:
|
|
from .mistral import *
|
|
except ImportError:
|
|
print("This version of transformers doesn't support mistral.")
|