From 43940ec67323e3370c2bad5fa11c3f76fc595fe1 Mon Sep 17 00:00:00 2001 From: Goekdeniz-Guelmez Date: Tue, 4 Feb 2025 11:13:07 +0100 Subject: [PATCH] fix Test --- llms/mlx_lm/lora.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/llms/mlx_lm/lora.py b/llms/mlx_lm/lora.py index 64be0a91..ff3ff752 100644 --- a/llms/mlx_lm/lora.py +++ b/llms/mlx_lm/lora.py @@ -287,16 +287,16 @@ def evaluate_model(args, model: nn.Module, tokenizer: TokenizerWrapper, test_set model.eval() if args.training_mode == "orpo": - test_loss, test_rewards = evaluate_orpo( + test_loss, test_rewards, _, _ = evaluate_orpo( model=model, dataset=test_set, - tokenizer=tokenizer, batch_size=args.batch_size, num_batches=args.test_batches, max_seq_length=args.max_seq_length, beta=args.beta ) - print(f"Test loss {test_loss:.3f}, Rewards: {test_rewards[0]:.3f}, {test_rewards[1]:.3f}") + test_ppl = math.exp(test_loss) + print(f"Test loss {test_loss:.3f}, Test ppl {test_ppl:.3f}, Rewards: {test_rewards[0]:.3f}, {test_rewards[1]:.3f}") else: test_loss = evaluate( model=model,