diff --git a/.github/workflows/run_chatgpt_examples.yml b/.github/workflows/run_chatgpt_examples.yml index e4af6ac6b..e269f392b 100644 --- a/.github/workflows/run_chatgpt_examples.yml +++ b/.github/workflows/run_chatgpt_examples.yml @@ -21,7 +21,7 @@ jobs: container: image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 options: --gpus all --rm -v /data/scratch/examples-data:/data/scratch/examples-data --shm-size=10.24gb - timeout-minutes: 60 + timeout-minutes: 180 defaults: run: shell: bash @@ -34,6 +34,10 @@ jobs: pip install --no-cache-dir -v -e . - name: Install ChatGPT + env: + CFLAGS: "-O1" + CXXFLAGS: "-O1" + MAX_JOBS: 4 run: | pip install flash-attn --no-build-isolation cd applications/ColossalChat diff --git a/.github/workflows/run_chatgpt_unit_tests.yml b/.github/workflows/run_chatgpt_unit_tests.yml index ef928c8dd..d8a1ce246 100644 --- a/.github/workflows/run_chatgpt_unit_tests.yml +++ b/.github/workflows/run_chatgpt_unit_tests.yml @@ -21,7 +21,7 @@ jobs: container: image: image-cloud.luchentech.com/hpcaitech/pytorch-cuda:2.2.2-12.1.0 options: --gpus all --rm -v /data/scratch/examples-data:/data/scratch/examples-data - timeout-minutes: 30 + timeout-minutes: 180 defaults: run: shell: bash @@ -30,6 +30,10 @@ jobs: uses: actions/checkout@v2 - name: Install ChatGPT + env: + CFLAGS: "-O1" + CXXFLAGS: "-O1" + MAX_JOBS: 4 run: | pip install flash-attn --no-build-isolation cd applications/ColossalChat diff --git a/applications/ColossalChat/coati/distributed/grpo_consumer.py b/applications/ColossalChat/coati/distributed/grpo_consumer.py index a2c3e03d6..424d46098 100644 --- a/applications/ColossalChat/coati/distributed/grpo_consumer.py +++ b/applications/ColossalChat/coati/distributed/grpo_consumer.py @@ -530,4 +530,4 @@ class GRPOConsumer(BaseConsumer): model = self.policy_model.unwrap() state_dict = model.state_dict() state_dict["consumer_global_step"] = torch.tensor([self.global_step], device=self.device) - return state_dict \ No newline at end of file + return state_dict diff --git a/colossalai/shardformer/modeling/qwen3.py b/colossalai/shardformer/modeling/qwen3.py index 437693800..5f96f5f49 100644 --- a/colossalai/shardformer/modeling/qwen3.py +++ b/colossalai/shardformer/modeling/qwen3.py @@ -273,7 +273,7 @@ class Qwen3PipelineForwards: hidden_states: Optional[torch.FloatTensor] = None, stage_index: Optional[List[int]] = None, shard_config: ShardConfig = None, - **kwargs + **kwargs, ): r""" Args: