Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Very slow alpaca-eval times for OLMo models #143

Open
dwadden opened this issue Apr 8, 2024 · 3 comments · May be fixed by #151
Open

Very slow alpaca-eval times for OLMo models #143

dwadden opened this issue Apr 8, 2024 · 3 comments · May be fixed by #151

Comments

@dwadden
Copy link
Contributor

dwadden commented Apr 8, 2024

Alpaca-eval on OLMo models is very slow -- maybe just because OLMo can't use vllm and huggingface in general is slow? Here's an example Beaker job; based on the TQDM log it will take 100 hours (~4 days) to evaluate 800 examples. This isn't really a workable solution. Potential options:

  • Get OLMo to play nicely with vllm. We'd probably need to recruit an engineer to help with this.
  • Evaluate on a subset of maybe 100 instances for OLMo models. Still not ideal but better than nothing.

@hamishivi @yizhongw any thoughts?

@hamishivi
Copy link
Collaborator

My understanding is that vLLM integration of OLMo was at some point being looked at by @AkshitaB , although not sure on the current setup.
As for the second, feel free to add a subset flag for this, since it might be something useful for debugging anyway. Just reducing the prompt set used should work naively with the existing code (I do this myself for debugging).

@AkshitaB
Copy link

AkshitaB commented Apr 9, 2024

@dwadden vllm supports OLMo in their latest version already, you should be able to use it directly. You'll need to convert the olmo checkpoint to HF format using the conversion script.

Also, make sure to use the latest vllm, since they fixed a bug with tensor parallel case in this commit after their last pip release.

@hamishivi
Copy link
Collaborator

thanks akshita!!!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants