diff --git a/configs/train/finetune.yaml b/configs/train/finetune.yaml index 9724bdbd..47b46f8e 100644 --- a/configs/train/finetune.yaml +++ b/configs/train/finetune.yaml @@ -2,27 +2,29 @@ model_name: "zpn/llama-7b" tokenizer_name: "zpn/llama-7b" gradient_checkpointing: true +save_name: "nomic-ai/vicuna-full-multi-turn" # dataset streaming: false num_proc: 64 -dataset_path: "data.jsonl" -max_length: 512 +dataset_path: "data_multiturn" +max_length: 1024 batch_size: 32 # train dynamics lr: 5.0e-5 -eval_every: 2000 +eval_every: 800 eval_steps: 100 -save_every: 2000 -output_dir: "ckpts/llama-7b" +save_every: 800 +output_dir: "ckpts/llama-7b-full-multi" checkpoint: null lora: false warmup_steps: 100 +num_epochs: 2 # logging -wandb: false -wandb_entity: zanussbaum -wandb_project: llama +wandb: true +wandb_entity: vicuna +wandb_project_name: vicuna seed: 42