-
Notifications
You must be signed in to change notification settings - Fork 2.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Phi3 lora module not loading #4915
Comments
yes, you can refer to #4715 for a conversation script. However phi3 128k contains also a gateup_proj, you can modify the script to decompose gate_up_proj layer as well. Like following:
It can load it without error, but not sure if there will be any performance issue. |
Thanks a lot. I will use this, but will VLLM be adding support for this in the future? |
After decomposing layers of Phi3, will there be any problem merging LoRA layers back into Phi3? Do we have to merge the decomposed layers back before append LoRA back onto the original Phi3? |
I added the new projection layers using the conversion code above and the standard vLLM code for LoRA using snapshot_download using their documentation (https://docs.vllm.ai/en/latest/models/lora.html#using-lora-adapters). Update: There was a bug in my code, the script above works just fine and the LoRa weights are merged well |
I guess they will, otherwise we will always need to convert the lora first, which is not very convenient. |
I think you can directly merge the original (not decomposed) LORA adapter to Phi3 if you're not trying to load LORA via vllm. |
ValueError: While loading /data/llm_resume_profiles_phi3_v1, expected target modules in ['q_proj', 'k_proj', 'v_proj', 'o_proj', 'gate_proj', 'up_proj', 'down_proj', 'embed_tokens', 'lm_head'] but received ['qkv_proj', 'gate_up_proj']. Please verify that the loaded LoRA module is correct.
I am unable to load LORA module with phi3-128k-instruct version. Can support for this be added?
I am using VLLM docker with version 0.4.2
thanks
The text was updated successfully, but these errors were encountered: