mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-11 22:10:37 +00:00
[Feature] auto-cast optimizers to distributed version (#5746)
* auto-cast optimizers to distributed * fix galore casting * logger --------- Co-authored-by: Edenzzzz <wtan45@wisc.edu>
This commit is contained in:
@@ -416,7 +416,7 @@ def exam_bert_test_on_hybrid_plugin(test_config):
|
||||
sharded_optimizer,
|
||||
criterion,
|
||||
booster,
|
||||
) = build_model_from_hybrid_plugin(model_fn, loss_fn, test_config, CAME, DistributedCAME)
|
||||
) = build_model_from_hybrid_plugin(model_fn, loss_fn, test_config, CAME, CAME)
|
||||
|
||||
org_loss, org_output, sharded_loss, sharded_output = run_forward_backward_with_hybrid_plugin(
|
||||
org_model, sharded_model, sharded_optimizer, data_gen_fn, output_transform_fn, criterion, booster
|
||||
|
Reference in New Issue
Block a user