fix CI machine tag

This commit is contained in:
BurkeHulk 2025-06-02 17:45:40 +08:00
parent fd56b22278
commit c9cba49ab5
19 changed files with 24 additions and 20 deletions

View File

@ -34,7 +34,7 @@ jobs:
anyExtensionFileChanged: ${{ steps.find-extension-change.outputs.any_changed }} anyExtensionFileChanged: ${{ steps.find-extension-change.outputs.any_changed }}
changedLibraryFiles: ${{ steps.find-lib-change.outputs.all_changed_files }} changedLibraryFiles: ${{ steps.find-lib-change.outputs.all_changed_files }}
anyLibraryFileChanged: ${{ steps.find-lib-change.outputs.any_changed }} anyLibraryFileChanged: ${{ steps.find-lib-change.outputs.any_changed }}
runs-on: [self-hosted,ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
concurrency: concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}-detect-change group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}-detect-change
cancel-in-progress: true cancel-in-progress: true
@ -87,7 +87,7 @@ jobs:
name: Build and Test Colossal-AI name: Build and Test Colossal-AI
needs: detect needs: detect
if: needs.detect.outputs.anyLibraryFileChanged == 'true' if: needs.detect.outputs.anyLibraryFileChanged == 'true'
runs-on: [self-hosted,ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
options: --gpus all --shm-size=2g --rm -v /dev/shm -v /data/scratch:/data/scratch options: --gpus all --shm-size=2g --rm -v /dev/shm -v /data/scratch:/data/scratch

View File

@ -10,7 +10,7 @@ jobs:
build: build:
name: Build and Test Colossal-AI name: Build and Test Colossal-AI
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
options: --gpus all --rm -v /dev/shm -v /data/scratch/:/data/scratch/ options: --gpus all --rm -v /dev/shm -v /data/scratch/:/data/scratch/

View File

@ -10,7 +10,7 @@ jobs:
matrix_preparation: matrix_preparation:
name: Prepare Container List name: Prepare Container List
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
runs-on: [self-hosted,ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
outputs: outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }} matrix: ${{ steps.set-matrix.outputs.matrix }}
steps: steps:

View File

@ -15,7 +15,7 @@ jobs:
if: | if: |
github.event.pull_request.draft == false && github.event.pull_request.draft == false &&
github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI' github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI'
runs-on: ubuntu-[self-hosted, ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
concurrency: concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}-check-i18n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}-check-i18n
cancel-in-progress: true cancel-in-progress: true

View File

@ -54,7 +54,7 @@ jobs:
needs.detect-changed-doc.outputs.any_changed == 'true' needs.detect-changed-doc.outputs.any_changed == 'true'
name: Test the changed Doc name: Test the changed Doc
needs: detect-changed-doc needs: detect-changed-doc
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
options: --gpus all --rm options: --gpus all --rm

View File

@ -10,7 +10,7 @@ jobs:
# Add this condition to avoid executing this job if the trigger event is workflow_dispatch. # Add this condition to avoid executing this job if the trigger event is workflow_dispatch.
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
name: Test the changed Doc name: Test the changed Doc
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
options: --gpus all --rm options: --gpus all --rm

View File

@ -40,7 +40,7 @@ jobs:
github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI' github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI'
name: Manually check example files name: Manually check example files
needs: manual_check_matrix_preparation needs: manual_check_matrix_preparation
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
strategy: strategy:
fail-fast: false fail-fast: false
matrix: ${{fromJson(needs.manual_check_matrix_preparation.outputs.matrix)}} matrix: ${{fromJson(needs.manual_check_matrix_preparation.outputs.matrix)}}

View File

@ -85,7 +85,7 @@ jobs:
needs.detect-changed-example.outputs.anyChanged == 'true' needs.detect-changed-example.outputs.anyChanged == 'true'
name: Test the changed example name: Test the changed example
needs: detect-changed-example needs: detect-changed-example
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
strategy: strategy:
fail-fast: false fail-fast: false
matrix: ${{fromJson(needs.detect-changed-example.outputs.matrix)}} matrix: ${{fromJson(needs.detect-changed-example.outputs.matrix)}}

View File

@ -10,7 +10,7 @@ jobs:
matrix_preparation: matrix_preparation:
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
name: Prepare matrix for weekly check name: Prepare matrix for weekly check
runs-on: ubunt[self-hosted, ubuntu-latest]u-latest runs-on: [self-hosted, ubuntu-latest]
outputs: outputs:
matrix: ${{ steps.setup-matrix.outputs.matrix }} matrix: ${{ steps.setup-matrix.outputs.matrix }}
steps: steps:
@ -29,7 +29,7 @@ jobs:
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
name: Weekly check all examples name: Weekly check all examples
needs: matrix_preparation needs: matrix_preparation
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
strategy: strategy:
fail-fast: false fail-fast: false
matrix: ${{fromJson(needs.matrix_preparation.outputs.matrix)}} matrix: ${{fromJson(needs.matrix_preparation.outputs.matrix)}}

View File

@ -9,7 +9,7 @@ jobs:
release: release:
name: Publish Docker Image to DockerHub name: Publish Docker Image to DockerHub
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: "hpcaitech/docker-in-docker:latest" image: "hpcaitech/docker-in-docker:latest"
options: --gpus all --rm -v /var/run/docker.sock:/var/run/docker.sock options: --gpus all --rm -v /var/run/docker.sock:/var/run/docker.sock

View File

@ -9,7 +9,7 @@ jobs:
publish: publish:
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
name: Build and publish Python 🐍 distributions 📦 to PyPI name: Build and publish Python 🐍 distributions 📦 to PyPI
runs-on: ubuntu-[self-hosted, ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
timeout-minutes: 20 timeout-minutes: 20
outputs: outputs:
status: ${{ steps.publish.outcome }} status: ${{ steps.publish.outcome }}

View File

@ -6,6 +6,10 @@ on:
- 'version.txt' - 'version.txt'
permissions:
id-token: write
contents: read
jobs: jobs:
build-n-publish: build-n-publish:
if: github.event_name == 'workflow_dispatch' || github.repository == 'hpcaitech/ColossalAI' if: github.event_name == 'workflow_dispatch' || github.repository == 'hpcaitech/ColossalAI'

View File

@ -10,7 +10,7 @@ jobs:
generate-and-publish: generate-and-publish:
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
name: Generate leaderboard report and publish to Lark name: Generate leaderboard report and publish to Lark
runs-on: ubuntu-[self-hosted, ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
timeout-minutes: 20 timeout-minutes: 20
steps: steps:
- uses: actions/checkout@v2 - uses: actions/checkout@v2

View File

@ -8,7 +8,7 @@ on:
jobs: jobs:
report-test-coverage: report-test-coverage:
runs-on: ubuntu-[self-hosted, ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
if: ${{ github.event.workflow_run.conclusion == 'success' }} if: ${{ github.event.workflow_run.conclusion == 'success' }}
steps: steps:
- name: "Download artifact" - name: "Download artifact"

View File

@ -17,7 +17,7 @@ jobs:
github.event.pull_request.draft == false && github.event.pull_request.draft == false &&
github.base_ref == 'main' && github.base_ref == 'main' &&
github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI' github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI'
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
options: --gpus all --rm -v /data/scratch/examples-data:/data/scratch/examples-data --shm-size=10.24gb options: --gpus all --rm -v /data/scratch/examples-data:/data/scratch/examples-data --shm-size=10.24gb

View File

@ -17,7 +17,7 @@ jobs:
github.event.pull_request.draft == false && github.event.pull_request.draft == false &&
github.base_ref == 'main' && github.base_ref == 'main' &&
github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI' github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI'
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
options: --gpus all --rm -v /data/scratch/examples-data:/data/scratch/examples-data options: --gpus all --rm -v /data/scratch/examples-data:/data/scratch/examples-data

View File

@ -17,7 +17,7 @@ jobs:
github.event.pull_request.draft == false && github.event.pull_request.draft == false &&
github.base_ref == 'main' && github.base_ref == 'main' &&
github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI' github.event.pull_request.base.repo.full_name == 'hpcaitech/ColossalAI'
runs-on: [self-hosted, gpu] runs-on: [self-hosted, ubuntu-latest]
container: container:
image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0
volumes: volumes:

View File

@ -7,7 +7,7 @@ on:
jobs: jobs:
sync-submodule: sync-submodule:
runs-on: ubuntu-[self-hosted, ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
if: github.repository == 'hpcaitech/ColossalAI' if: github.repository == 'hpcaitech/ColossalAI'
steps: steps:
- name: Checkout - name: Checkout

View File

@ -7,7 +7,7 @@ on:
jobs: jobs:
build: build:
runs-on: ubuntu-[self-hosted, ubuntu-latest] runs-on: [self-hosted, ubuntu-latest]
steps: steps:
- uses: usthe/issues-translate-action@v2.7 - uses: usthe/issues-translate-action@v2.7
with: with: