Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: gguf loading failed. config.json? #417

Open
juud79 opened this issue Apr 17, 2024 · 4 comments
Open

[Bug]: gguf loading failed. config.json? #417

juud79 opened this issue Apr 17, 2024 · 4 comments
Labels
bug Something isn't working

Comments

@juud79
Copy link

juud79 commented Apr 17, 2024

Your current environment

I excute that command below:

python -m aphrodite.endpoints.openai.api_server --model /root/.cache/huggingface/hub/gguf/ --quantization gguf --gpu-memory-utilization 0.35 --max-model-len 4096 --port 8000

there is an error that
OSError: /root/.cache/huggingface/hub/gguf/ does not appear to have a file named config.json. Checkout 'https://huggingface.co//root/.cache/huggingface/hub/gguf//tree/None' for available files.

why there wants config.json?
as you know gguf format doesn't have a config.json...

🐛 Describe the bug

I excute that command below:

python -m aphrodite.endpoints.openai.api_server --model /root/.cache/huggingface/hub/gguf/ --quantization gguf --gpu-memory-utilization 0.35 --max-model-len 4096 --port 8000

there is an error that
OSError: /root/.cache/huggingface/hub/gguf/ does not appear to have a file named config.json. Checkout 'https://huggingface.co//root/.cache/huggingface/hub/gguf//tree/None' for available files.

why there wants config.json?
as you know gguf format doesn't have a config.json...

@juud79 juud79 added the bug Something isn't working label Apr 17, 2024
@sgsdxzy
Copy link
Collaborator

sgsdxzy commented Apr 17, 2024

You need to point to the file (xxxx.gguf), not the directory containing the file.

@juud79
Copy link
Author

juud79 commented Apr 17, 2024

You need to point to the file (xxxx.gguf), not the directory containing the file.

model is consist
of 2 gguf files.. how can i do that?

@sgsdxzy
Copy link
Collaborator

sgsdxzy commented Apr 23, 2024

sharded gguf (a model in multiple files) is not currently supported. #420 adds support but we need to fix something else related to ggufs first.

@sgsdxzy
Copy link
Collaborator

sgsdxzy commented Apr 30, 2024

Experimental support of multiple gguf files is added to the dev branch, please test if it works according to the documentation

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants