mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-07-30 23:06:01 +00:00
[CI] use flash_attn==2.7.4.post1
This commit is contained in:
parent
4b7727996e
commit
12c55538f5
4
.github/workflows/build_on_pr.yml
vendored
4
.github/workflows/build_on_pr.yml
vendored
@ -138,6 +138,10 @@ jobs:
|
||||
cp -p -r /github/home/cuda_ext_cache/* /__w/ColossalAI/ColossalAI/
|
||||
fi
|
||||
|
||||
- name: Install flash-attention
|
||||
run: |
|
||||
pip install flash-attn==2.7.4.post1 --no-build-isolation
|
||||
|
||||
- name: Install Colossal-AI
|
||||
run: |
|
||||
BUILD_EXT=1 pip install -v -e .
|
||||
|
Loading…
Reference in New Issue
Block a user