-
Notifications
You must be signed in to change notification settings - Fork 166
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
基于上次提的问题#691,后续改进后似乎依旧不能按微调的情况回复。 #880
Comments
不merge就没问题?merge了就不太行? |
是的,之前是使用看量化微调,所以不行,可以理解,但后续取消了量化微调,merge了依旧不行?着实有点想不明白。 为了方便你查看,我将把我merge过程中的log输出展示出来:“
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
基于上次提的问题#691,后续改进后似乎依旧不能按微调的情况回复。
您好,根据上次的反馈,我在这次使用的微调中,直接使用lora微调,并没有使用量化。quantization_bit = 0;
训练模型后,将模型进行合并(为了后续可以使用VLLM进行推理。)
此时发现,这样做,模型依旧未能按照实际微调的结果(即未合并前的模型)来进行回复。
想请问老师,到底时哪一步出现问题了?很是奇怪。
The text was updated successfully, but these errors were encountered: