mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-10-28 20:30:42 +00:00
[example] integrate seq-parallel tutorial with CI (#2463)
This commit is contained in:
@@ -1,11 +1,8 @@
|
||||
from colossalai.amp import AMP_TYPE
|
||||
|
||||
DATA_PATH = ''
|
||||
VOCAB_FILE_PATH = ''
|
||||
|
||||
# hyper-parameters
|
||||
TRAIN_ITERS = 1000000
|
||||
DECAY_ITERS = 990000
|
||||
TRAIN_ITERS = 10
|
||||
DECAY_ITERS = 4
|
||||
WARMUP_FRACTION = 0.01
|
||||
GLOBAL_BATCH_SIZE = 32 # dp world size * sentences per GPU
|
||||
EVAL_ITERS = 10
|
||||
@@ -13,12 +10,12 @@ EVAL_INTERVAL = 10
|
||||
LR = 0.0001
|
||||
MIN_LR = 1e-05
|
||||
WEIGHT_DECAY = 0.01
|
||||
SEQ_LENGTH = 512
|
||||
SEQ_LENGTH = 128
|
||||
|
||||
# BERT config
|
||||
DEPTH = 12
|
||||
NUM_ATTENTION_HEADS = 12
|
||||
HIDDEN_SIZE = 768
|
||||
DEPTH = 4
|
||||
NUM_ATTENTION_HEADS = 4
|
||||
HIDDEN_SIZE = 128
|
||||
|
||||
# model config
|
||||
ADD_BINARY_HEAD = False
|
||||
|
||||
Reference in New Issue
Block a user