mirror of
				https://github.com/hpcaitech/ColossalAI.git
				synced 2025-11-03 23:48:41 +00:00 
			
		
		
		
	Benchmarks
Benchmark OPT with LoRA on dummy prompt data
We provide various OPT models (string in parentheses is the corresponding model name used in this script):
- OPT-125M (125m)
 - OPT-350M (350m)
 - OPT-700M (700m)
 - OPT-1.3B (1.3b)
 - OPT-2.7B (2.7b)
 - OPT-3.5B (3.5b)
 - OPT-5.5B (5.5b)
 - OPT-6.7B (6.7b)
 - OPT-10B (10b)
 - OPT-13B (13b)
 
We also provide various training strategies:
- gemini: ColossalAI GeminiPlugin with 
placement_policy="cuda", like zero3 - gemini_auto: ColossalAI GeminiPlugin with 
placement_policy="cpu", like zero3-offload - zero2: ColossalAI zero2
 - zero2_cpu: ColossalAI zero2-offload
 - 3d: ColossalAI HybridParallelPlugin with TP, DP support
 
How to Run
cd ../tests
# Prepare data for benchmark
SFT_DATASET=/path/to/sft/data/ \
PROMPT_DATASET=/path/to/prompt/data/ \
PRETRAIN_DATASET=/path/to/ptx/data/ \
PREFERENCE_DATASET=/path/to/preference/data \
./test_data_preparation.sh
# Start benchmark
./benchmark_ppo.sh