[chatgpt]fix train_rm bug with lora (#2741)

This commit is contained in:
BlueRum
2023-02-16 10:25:17 +08:00
committed by GitHub
parent b6e3b955c3
commit 648183a960

View File

@@ -43,7 +43,7 @@ class RewardModelTrainer(ABC):
# train
if use_lora > 0:
print("Using Lora")
lora.mark_only_lora_as_trainable(self.model)
lora.mark_only_lora_as_trainable(self.model.model)
else:
self.model.train()
for chosen_ids, c_mask, reject_ids, r_mask in self.train_dataloader: