mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-15 22:19:38 +00:00
[pipeline] rewrite t5 tests & support multi-tensor transmitting in pipeline (#4388)
* fix remaining t5 bugs/rewrite t5 tests * fix multi-tensor communication in pipeline * rearrange test_config * fix keyerror in sync_shared_params * fix get_held_layers & Randomnizer, complete t5 tests * erase printing * fix get_held_layers through modifying _release_unheld_layers * fix _get_recursive_held_layers bug
This commit is contained in:
committed by
Hongxin Liu
parent
906426cb44
commit
ed4c448488
@@ -107,8 +107,15 @@ class OneForwardOneBackwardSchedule(PipelineSchedule):
|
||||
if output_obj_grad is None:
|
||||
optimizer.backward(output_obj)
|
||||
else:
|
||||
for k, grad in output_obj_grad.items():
|
||||
optimizer.backward_by_grad(output_obj[k], grad)
|
||||
if "backward_tensor_keys" not in output_obj:
|
||||
for k, grad in output_obj_grad.items():
|
||||
optimizer.backward_by_grad(output_obj[k], grad)
|
||||
else:
|
||||
for k, grad in output_obj_grad.items():
|
||||
output_obj[k].grad = grad
|
||||
for k in output_obj["backward_tensor_keys"]:
|
||||
tensor_to_backward = output_obj[k]
|
||||
optimizer.backward_by_grad(tensor_to_backward, tensor_to_backward.grad)
|
||||
|
||||
# Collect the grad of the input_obj.
|
||||
input_obj_grad = None
|
||||
|
Reference in New Issue
Block a user