-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Number of tokens (757) exceeded maximum context length (512). #2
Comments
i got the just same as @datacrud8 . did any one got solved ? thanks in advance. $ chainlit run main.py -w |
hello, can you try a different embeddings model, for example, |
@sudarshan-koirala At first, thanks for the answer to my question. into below lists including yours recommend, from by refering (https://www.sbert.net/docs/pretrained_models.html#multi-lingual-models ) .
but got no luck ...
$ time python ingest.py real 2m12.113s |
As per new updates, define like this:- llm = CTransformers( |
hi, trying to build this app in local, and used same model llama-2-7b-chat.ggmlv3.q8_0.bin
when run the app UI showing some random message same like you showed but checking in console getting this below message:
Number of tokens (755) exceeded maximum context length (512).
Number of tokens (756) exceeded maximum context length (512).
Number of tokens (757) exceeded maximum context length (512).
so increased max_new_tokens=2048, and increased n_ctx and added truncate=True , non of them are fixing this issue.
Changed the model as well. still same issue.
do you know any solution for this issue?
The text was updated successfully, but these errors were encountered: