From 53bb3bcc0af8b5b7ef0fe0d13bda7d853e4ec78c Mon Sep 17 00:00:00 2001 From: "Kai Wang (Victor Kai)" <37533040+kaiwang960112@users.noreply.github.com> Date: Thu, 10 Mar 2022 11:33:21 +0800 Subject: [PATCH] fix format (#362) --- colossalai/amp/naive_amp/_fp16_optimizer.py | 2 +- colossalai/nn/optimizer/cpu_adam.py | 6 ------ 2 files changed, 1 insertion(+), 7 deletions(-) diff --git a/colossalai/amp/naive_amp/_fp16_optimizer.py b/colossalai/amp/naive_amp/_fp16_optimizer.py index b1fc621c2..01842590f 100644 --- a/colossalai/amp/naive_amp/_fp16_optimizer.py +++ b/colossalai/amp/naive_amp/_fp16_optimizer.py @@ -14,7 +14,7 @@ from colossalai.context.parallel_mode import ParallelMode from colossalai.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.utils import (print_rank_0, copy_tensor_parallel_attributes, - clip_grad_norm_fp32, count_zeros_fp32, multi_tensor_applier, is_using_pp) + clip_grad_norm_fp32, count_zeros_fp32, multi_tensor_applier) def _zero_grad_group_helper(group, set_to_none): diff --git a/colossalai/nn/optimizer/cpu_adam.py b/colossalai/nn/optimizer/cpu_adam.py index 21f607c48..55f50bfdc 100644 --- a/colossalai/nn/optimizer/cpu_adam.py +++ b/colossalai/nn/optimizer/cpu_adam.py @@ -1,10 +1,4 @@ -# modified from https://github.com/microsoft/DeepSpeed/blob/master/deepspeed/ops/adam/cpu_adam.py - -import math import torch -import time -from pathlib import Path -import colossalai class CPUAdam(torch.optim.Optimizer):