Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: KV Cache and Max Tokens - Lack of Consistency #362

Open
official-elinas opened this issue Mar 28, 2024 · 0 comments
Open

[Bug]: KV Cache and Max Tokens - Lack of Consistency #362

official-elinas opened this issue Mar 28, 2024 · 0 comments
Labels
bug Something isn't working

Comments

@official-elinas
Copy link
Contributor

Your current environment

SW/Hardware;
Kubuntu 23.10
EPYC 7B13
2x 3090s
256GB DDR4 ECC

馃悰 Describe the bug

Running Aphrodite as I normally do, for example:

./runtime.sh python -m aphrodite.endpoints.openai.api_server --model /media/x/nvme_1/models/Midnight-Miqu-70B-v1.5_GPTQ32G -q gptq --dtype float16 -tp 2 --port 5069 --host "192.168.1.6" --tokenizer-mode auto --max-model-len 16000 -gmu .97 --disable-log-requests --served-model-name "sambarnes/Midnight-Miqu-70B-v1.5_GPTQ32G" --enforce-eager --kv-cache-dtype fp8_e5m2

Then I get different max tokens that can be stored in the KV cache, this one being 96... the lowest I've seen.

ValueError: The model's max seq len (16000) is larger than the maximum number of tokens that can be stored in KV cache (96). Try increasing 'gpu_memory_utilization' or decreasing 'max_model_len' when initializing the engine

I brought this up in the KoboldAI discord before in the #aphrodite channel. I've seen as high as 48K and have been able to run at 32K many times.

I am on the latest commit and this has been happening for a bit, similar to a while back (months ago) when ray would fail for multiple launches.

@official-elinas official-elinas added the bug Something isn't working label Mar 28, 2024
@official-elinas official-elinas changed the title [Bug]: KV Cache and Max Token - Lack of Consistency [Bug]: KV Cache and Max Tokens - Lack of Consistency Mar 28, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant