-
Notifications
You must be signed in to change notification settings - Fork 46
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QA] Where to download DeepSeek-R1 gptq model? #1267
Comments
You can visit https://huggingface.co/models?search=gptq to download our DeepSeek R1 distlled 7B model but we currently do not provide the full R1 model. You can use our toolkit to quantize your down R1 model. |
Deepseek.ai has released the FP8 version. Can our tools work with it directly? |
You can use the bf16 version of R1 to GPTQ quantize. We do not have large H100+ gpu to test FP8 model load. 4090 has too little vram. |
Great, Thanks! |
One more question, have you tested if there are any issues with DeepSeek R1 GPTQ inference? Can it be used for inference with the |
There are no technical reasons why GPTQ quantized R1 cannot run on vLLM or SGLang. |
Hello, I am quite interested in your work. I would like to ask you a few questions:
|
Could you please tell me which deepseek-7B model you can compress? If convenient, please provide the link of 7B model. |
|
https://arxiv.org/abs/2210.17323 |
|
This paper was written by the original researchers of GPTQ. GPTQModel is code, based on the original code from the original research team plus many modifications on usage, inference, and quantization. |
How to Download DeepSeek-v1 gptq quanted model?
The text was updated successfully, but these errors were encountered: