-
Notifications
You must be signed in to change notification settings - Fork 500
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AttributeError: 'Qwen2Model' object has no attribute 'lm_head' #500
Comments
This error also occurs on some Llama models:
In this case, it was a Llama 3.2 1B finetine: huihui-ai_Llama-3.2-1B-Instruct-abliterated |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I get this error when trying to extract loras from Qwen models. It doesn't happen for all qwen models, but it seems to mostly be the 0.5/1.5/3B models and I cannot figure out what causes it, or why 7/14/32B succeed without issue. This particular example is from "mobiuslabsgmbh_DeepSeek-R1-ReDistill-Qwen-1.5B-v1.1", while some others like "huihui-ai_DeepSeek-R1-Distill-Qwen-7B-abliterated" work just fine. I am running the latest version of mergekit, and I have the latest transformers / torch installed, as far as I can tell.
Any clues as to what the problem is would be really helpful. Thanks!
The text was updated successfully, but these errors were encountered: