Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama 3 8B GPU requirements? #979

Closed
Leonard907 opened this issue May 14, 2024 · 2 comments
Closed

Llama 3 8B GPU requirements? #979

Leonard907 opened this issue May 14, 2024 · 2 comments

Comments

@Leonard907
Copy link

Thanks for the integration for Llama 3 models. I'm interested in fine-tuning Llama 3 8B with full 8K context using LoRA. What resource requirements are needed? Can I do it in just 1 A100 80GB GPU?

@ebsmothers
Copy link
Contributor

Hi @Leonard907 thanks for creating the issue! The exact resource requirements depend on batch size and other training configs, but if I run for instance

tune run lora_finetune_single_device --config llama3/8B_lora_single_device \
dataset=torchtune.datasets.slimorca_dataset dataset.max_seq_len=8192

I am able to fit pretty comfortably into a single A100 80GB.

@Leonard907
Copy link
Author

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants