From a5883aa7909070480d218b62ff8f3e987e7eebd8 Mon Sep 17 00:00:00 2001 From: Frank Lee Date: Fri, 16 Jun 2023 18:23:02 +0800 Subject: [PATCH] [test] fixed codefactor format report (#4026) --- .../test_general_checkpoint_io.py | 18 ++++++++++++++---- 1 file changed, 14 insertions(+), 4 deletions(-) diff --git a/tests/test_checkpoint_io/test_general_checkpoint_io.py b/tests/test_checkpoint_io/test_general_checkpoint_io.py index 88e3673c1..0976d4503 100644 --- a/tests/test_checkpoint_io/test_general_checkpoint_io.py +++ b/tests/test_checkpoint_io/test_general_checkpoint_io.py @@ -165,8 +165,13 @@ def test_sharded_optimizer_multiple_param_groups(): # create a model and optimizer model = resnet18() - optimizer = Adam([{'params': model.layer1.parameters()}, \ - {'params': model.layer2.parameters(), 'lr': 0.002}], lr=0.001) + optimizer = Adam([{ + 'params': model.layer1.parameters() + }, { + 'params': model.layer2.parameters(), + 'lr': 0.002 + }], + lr=0.001) # create test data sample x = torch.randn(1, 3, 224, 224) @@ -189,8 +194,13 @@ def test_sharded_optimizer_multiple_param_groups(): # create new model new_model = resnet18() - new_optimizer = Adam([{'params': new_model.layer1.parameters()}, \ - {'params': new_model.layer2.parameters(), 'lr': 0.002}], lr=0.001) + new_optimizer = Adam([{ + 'params': new_model.layer1.parameters() + }, { + 'params': new_model.layer2.parameters(), + 'lr': 0.002 + }], + lr=0.001) ckpt_io.load_model(new_model, str(model_ckpt_dir.name), strict=True) ckpt_io.load_optimizer(new_optimizer, str(optimizer_ckpt_dir.name))