[pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci
This commit is contained in:
pre-commit-ci[bot]
2025-08-14 11:05:40 +00:00
parent 99ba48fc40
commit 73bdfd8891
5 changed files with 10 additions and 10 deletions

View File

@@ -89,9 +89,7 @@ def train(args):
actor = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
if args.rm_pretrain:
reward_model = RewardModel(args.rm_pretrain, trust_remote_code=True)
ref_model = AutoModelForCausalLM.from_pretrained(
args.pretrain, trust_remote_code=True
)
ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
if args.lora_config is not None:
actor = convert_to_lora_module(actor, lora_config=lora_config)

View File

@@ -102,9 +102,7 @@ def train(args):
coordinator.print_on_master(msg="Flash-attention enabled successfully")
else:
actor = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
ref_model = AutoModelForCausalLM.from_pretrained(
args.pretrain, trust_remote_code=True
)
ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True)
if not args.no_neural_reward_model:
reward_model = RewardModel(args.rm_pretrain, trust_remote_code=True)
critic = Critic(args.rm_pretrain)