mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2026-04-11 14:43:10 +00:00
[pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
This commit is contained in:
@@ -89,9 +89,7 @@ def train(args):
|
||||
actor = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
|
||||
if args.rm_pretrain:
|
||||
reward_model = RewardModel(args.rm_pretrain, trust_remote_code=True)
|
||||
ref_model = AutoModelForCausalLM.from_pretrained(
|
||||
args.pretrain, trust_remote_code=True
|
||||
)
|
||||
ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
|
||||
|
||||
if args.lora_config is not None:
|
||||
actor = convert_to_lora_module(actor, lora_config=lora_config)
|
||||
|
||||
@@ -102,9 +102,7 @@ def train(args):
|
||||
coordinator.print_on_master(msg="Flash-attention enabled successfully")
|
||||
else:
|
||||
actor = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
|
||||
ref_model = AutoModelForCausalLM.from_pretrained(
|
||||
args.pretrain, trust_remote_code=True
|
||||
)
|
||||
ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
|
||||
if not args.no_neural_reward_model:
|
||||
reward_model = RewardModel(args.rm_pretrain, trust_remote_code=True)
|
||||
critic = Critic(args.rm_pretrain)
|
||||
|
||||
Reference in New Issue
Block a user