Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

finetuned using lamma-13B #13

Open
Huangbukun opened this issue Dec 22, 2023 · 3 comments
Open

finetuned using lamma-13B #13

Huangbukun opened this issue Dec 22, 2023 · 3 comments

Comments

@Huangbukun
Copy link

Hello, if I want to use llama-13B's pth for fine-tuning, what changes need to be made to the train.sh script? After fine-tuning according to the parameters of llama-7B, the accuracy is very low.

@ikodoh
Copy link
Contributor

ikodoh commented Dec 26, 2023

In our experiments, I changed --adapter_layer from 32 to 40, and you may also decrease the learning rate.

@Huangbukun
Copy link
Author

In our experiments, I changed --adapter_layer from 32 to 40, and you may also decrease the learning rate.

thank you!

@Huangbukun
Copy link
Author

In our experiments, I changed --adapter_layer from 32 to 40, and you may also decrease the learning rate.

Hello, after I adjusted --adapter_layer to 40, I changed the learning rate to 9e-3, but your result in the double line can only reach 65%.I don't know what I did wrong

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants