mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-25 03:31:56 +00:00
[optimizer] add div_scale for optimizers (#2117)
* [optimizer] add div_scale for optimizers * [zero] use div_scale in zero optimizer * fix testing error
This commit is contained in:
@@ -71,7 +71,7 @@ def test_adam(adamw, step, p_dtype, g_dtype):
|
||||
weight_decay = 0
|
||||
|
||||
multi_tensor_applier(fused_adam, dummy_overflow_buf, [[g], [p], [m], [v]], lr, beta1, beta2, eps, step, adamw,
|
||||
True, weight_decay)
|
||||
True, weight_decay, -1)
|
||||
|
||||
torch_adam_update(
|
||||
step,
|
||||
|
Reference in New Issue
Block a user