-
Notifications
You must be signed in to change notification settings - Fork 277
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the current text generation call will exceed the model's predefined maximum length (4096) #151
Comments
Same question here. |
I reset the config, but it still had the warning"the current text generation call will exceed the model's predefined maximum length (4096)" |
I also did and it still shows this warning. But I think it's not an error, it's just a warning. Also from my experimental results, my model can correctly understand and generate output with 32k input even with the warning showing up. |
I tested with a long text, and it indeed behaves as you described. Thank you for your response. |
Thanks for providing this project.
I am using interface.py to infer the model Llama-2-13b-chat-longlora-32k-sft. I input a text of 32,000 tokens and received the following warning:
here is my input:
I don't know how to solve this warning. If I run LongAlpaca-7B, it has the same warning.
The text was updated successfully, but these errors were encountered: