From e397327de01b675c49c3e95711b77579de040a62 Mon Sep 17 00:00:00 2001 From: "pre-commit-ci[bot]" <66853113+pre-commit-ci[bot]@users.noreply.github.com> Date: Wed, 16 Apr 2025 10:00:10 +0000 Subject: [PATCH] [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --- .../ColossalChat/coati/distributed/grpo_consumer.py | 9 +++++++-- 1 file changed, 7 insertions(+), 2 deletions(-) diff --git a/applications/ColossalChat/coati/distributed/grpo_consumer.py b/applications/ColossalChat/coati/distributed/grpo_consumer.py index c0cb7663b..8acf41807 100644 --- a/applications/ColossalChat/coati/distributed/grpo_consumer.py +++ b/applications/ColossalChat/coati/distributed/grpo_consumer.py @@ -1,7 +1,7 @@ import json import os from contextlib import nullcontext -from typing import Optional, Any +from typing import Any, Optional import ray import torch @@ -212,7 +212,12 @@ class GRPOConsumer(BaseConsumer): # update gradient only if at least 0.7*batch_size*num_generation valid samples are collected in case a lot of samples are invalid and got filtered out. # balance between efficiency and accuracy need_update = self.effective_sample_count >= self.batch_size * self.dp_size * self.num_generations * 0.75 - pbar.set_postfix({"Step": self.global_step + 1, "Status": f"Collecting: {self.effective_sample_count}/{self.batch_size * self.dp_size * self.num_generations * 0.75}"}) + pbar.set_postfix( + { + "Step": self.global_step + 1, + "Status": f"Collecting: {self.effective_sample_count}/{self.batch_size * self.dp_size * self.num_generations * 0.75}", + } + ) # Gradient must be synchronized if zero2 is enabled. https://github.com/hpcaitech/ColossalAI/blob/44d4053fec005fe0b06b6bc755fdc962463145df/colossalai/booster/plugin/hybrid_parallel_plugin.py#L1500 ctx = (