mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-08-11 21:01:54 +00:00
[CI] use flash_attn==2.7.4.post1
This commit is contained in:
parent
4b7727996e
commit
12c55538f5
4
.github/workflows/build_on_pr.yml
vendored
4
.github/workflows/build_on_pr.yml
vendored
@ -138,6 +138,10 @@ jobs:
|
|||||||
cp -p -r /github/home/cuda_ext_cache/* /__w/ColossalAI/ColossalAI/
|
cp -p -r /github/home/cuda_ext_cache/* /__w/ColossalAI/ColossalAI/
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
- name: Install flash-attention
|
||||||
|
run: |
|
||||||
|
pip install flash-attn==2.7.4.post1 --no-build-isolation
|
||||||
|
|
||||||
- name: Install Colossal-AI
|
- name: Install Colossal-AI
|
||||||
run: |
|
run: |
|
||||||
BUILD_EXT=1 pip install -v -e .
|
BUILD_EXT=1 pip install -v -e .
|
||||||
|
Loading…
Reference in New Issue
Block a user