Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Using a simple model for torchtune setup testing in PRs? #978

Open
mikekgfb opened this issue May 14, 2024 · 1 comment
Open

Using a simple model for torchtune setup testing in PRs? #978

mikekgfb opened this issue May 14, 2024 · 1 comment

Comments

@mikekgfb
Copy link

I'm trying to run torchtune as an on-pr test for torchchat to ensure on-going comptaibility. Alas, llama3 can't be used because on-pr tests can't use HF tokens. Can you please suggest an alternative setup, or enable stories15M from Andrej's tinyllamas collection as a standin?

https://github.com/pytorch/torchchat/actions/runs/9071903213/job/24926494708

  + tune download stories15M --output-dir ./Meta-Llama-3-8B
  Ignoring files matching the following patterns: *.safetensors
  usage: tune download <repo-id> [OPTIONS]
  tune download: error: Repository 'stories15M' not found on the Hugging Face Hub.

cc: @byjlw

@ebsmothers
Copy link
Contributor

Hi @mikekgfb one suggestion is to use one of the checkpoints here which you can just get via curl. These are more or less random weights though, so if you want to test specifically for generation quality you may want to go another route.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants