You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Llama3 keeps generating tokens until max_new_tokens. It looks like the eos_token_id is never registered.
I had similar issue running locally, and updating transformers to >4.40 solved the issue. Issue related to llama3b and llama3b-instruct having different eos_tokens.
System Info
lorax-client==0.5.0
Information
Tasks
Reproduction
.
Expected behavior
I use below code to get LLM response.
Llama3 keeps generating tokens until max_new_tokens. It looks like the eos_token_id is never registered.
I had similar issue running locally, and updating transformers to >4.40 solved the issue. Issue related to
llama3b
andllama3b-instruct
having differenteos_tokens
.I tried setting
stop_sequence
but this returns empty string response. What is the proper way to set stopping tokens?
Am I setting up Predibase correctly?
The text was updated successfully, but these errors were encountered: