Replies: 1 comment 2 replies
-
@ee-lang do you mind mentioning which parameters you are setting (repetition_penalty, top_k, top_p, max_tokens, temperature etc.), and their exact values? The responses that I'm getting are namely widely different. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I set the parameters to be identical (and of course temperature to be 0 and the same seed), but I get slightly different responses to exactly the same prompt when running llama.cpp vs llama_cpp.py.
Is there a difference in sampling? (i.e. in logits/probabilities calculation)
Beta Was this translation helpful? Give feedback.
All reactions