Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Setting seed to -1 (random) or using default LLAMA_DEFAULT_SEED generates a deterministic reply chain #1809

Open
m-from-space opened this issue Oct 24, 2024 · 1 comment

Comments

@m-from-space
Copy link

m-from-space commented Oct 24, 2024

INB4 This is not about setting Top-P to 1.0 which causes the same output every time for every same prompt, documented here: #1797

When loading a model with release 0.3.1 and set the seed to either -1 (random) or leave the default (which is supposed to use a RNG according to the docs), the first reply of a model will always be the same for the same prompt. Consecutive replies for the same prompt will be different, but the chain will stay the same for each time loading the model and repeating the steps.

This points towards the seed not being randomized on load.

What I expect and what worked earlier: Loading a model with a random seed will generate a different first reply for the same prompt.

The issue is not present in version llama-cpp-python==0.2.90

@m-from-space
Copy link
Author

Problem is still present in llama-cpp-python==0.3.2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant