Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: Fix Context length bug in local LLMs proposal #825

Open
Wladastic opened this issue Mar 30, 2024 · 4 comments
Open

[Bug]: Fix Context length bug in local LLMs proposal #825

Wladastic opened this issue Mar 30, 2024 · 4 comments
Labels
bug Something isn't working

Comments

@Wladastic
Copy link

Version

VisualStudio Code extension

Operating System

Windows 10

What happened?

image I found a potential fix for the context length bug where the llm keeps outputting gibberish. It still does so but does not get stuck on it now. Also increase the alpha_value to 3 and experiment with different n_batch values, like 1024 instead of 512 (increases the input length of context)
@Wladastic Wladastic added the bug Something isn't working label Mar 30, 2024
@Wladastic
Copy link
Author

Version

VisualStudio Code extension

Operating System

Windows 10

What happened?

image I found a potential fix for the context length bug where the llm keeps outputting gibberish. It still does so but does not get stuck on it now. Also increase the alpha_value to 3 and experiment with different n_batch values, like 1024 instead of 512 (increases the input length of context)

This is in llm_connection.py btw.

@Wladastic
Copy link
Author

Wladastic commented Mar 30, 2024

I also added this here, works like a charm to avoid gibberish:
data["truncation_length"] = MAX_GPT_MODEL_TOKENS * 2

Although it may make the output sometimes not as accurate as GPT-4 Turbo:
It does at least not fix the infinite "_A_A_A_A" and random token outputs finally.

@techjeylabs
Copy link
Collaborator

seems solved, therefore closing

@Wladastic
Copy link
Author

No its not solved, the code wasn’t fixed.
I can prepare a merge request later

@techjeylabs techjeylabs reopened this Apr 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants