Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproducing sentiment finetuning train_lora extremely slow #146

Open
vikigenius opened this issue Jan 5, 2024 · 1 comment
Open

Reproducing sentiment finetuning train_lora extremely slow #146

vikigenius opened this issue Jan 5, 2024 · 1 comment
Labels
help wanted Extra attention is needed

Comments

@vikigenius
Copy link

I am trying to reproduce the finetuning for the fingpt-sentiment_llama2-13b_lora

The table claims we can do this in just a single RTX 3090 within a day.
I am using a L4 GPU instead.

I downloaded the models to base_models and the dataset to data correctly

I used the script like this

deepspeed -i train_lora.py \
--run_name sentiment-llama2-13b-20epoch-64batch \
--base_model llama2-13b-nr \
--dataset sentiment-train \
--max_length 512 \
--batch_size 16 \
--learning_rate 1e-4 \
--num_epochs 20 \

I got an OOM.

So i set the load_in_8_bit=True

But I am getting extremely slow fine tuning speed A single epoch is estimated to take 2 days.

@zhumingpassional zhumingpassional added bug Something isn't working help wanted Extra attention is needed and removed bug Something isn't working labels Jan 30, 2024
@ynjiun
Copy link

ynjiun commented Feb 13, 2024

two things you might want to consider to speed up:

  1. --base_model llama2-13b-nr => --base_model llama2-7b-nr
  2. use rtx 3090 which is faster than L4 in in memory bandwidth and more cores

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

3 participants