[feat] support qwen3 in shardformer

This commit is contained in:
botbw
2025-07-10 13:57:52 +08:00
committed by GitHub
parent 97f4bee9d8
commit d097224d90
9 changed files with 1829 additions and 42 deletions

View File

@@ -138,6 +138,10 @@ jobs:
cp -p -r /github/home/cuda_ext_cache/* /__w/ColossalAI/ColossalAI/
fi
- name: Install flash-attention
run: |
pip install flash-attn==2.7.4.post1 --no-build-isolation
- name: Install Colossal-AI
run: |
BUILD_EXT=1 pip install -v -e .