mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2026-01-29 21:49:54 +00:00
* Add gradient accumulation, fix lr scheduler
* fix FP16 optimizer and adapted torch amp with tensor parallel (#18)
* fixed bugs in compatibility between torch amp and tensor parallel and performed some minor fixes
* fixed trainer
* Revert "fixed trainer"
This reverts commit 2e0b0b7699.
* improved consistency between trainer, engine and schedule (#23)
Co-authored-by: 1SAA <c2h214748@gmail.com>
Co-authored-by: 1SAA <c2h214748@gmail.com>
Co-authored-by: ver217 <lhx0217@gmail.com>
23 lines
304 B
Python
23 lines
304 B
Python
#!/usr/bin/env python
|
|
# -*- encoding: utf-8 -*-
|
|
|
|
NUM_EPOCH = int
|
|
|
|
model = dict()
|
|
train_data = dict()
|
|
test_data = dict()
|
|
optimizer = dict()
|
|
loss = dict()
|
|
|
|
fp16 = dict()
|
|
zero = dict()
|
|
|
|
gradient_handler = []
|
|
parallel = dict()
|
|
hooks = []
|
|
|
|
cudnn_benchmark = True
|
|
cudnn_deterministic = False
|
|
|
|
logging = dict()
|