-
Notifications
You must be signed in to change notification settings - Fork 25
Issues: vllm-project/vllm-ascend
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Add a turtorial for Qwen 2.5-VL
documentation
Improvements or additions to documentation
help wanted
Extra attention is needed
#75
opened Feb 17, 2025 by
Yikun
running of v0.7.1 problem
documentation
Improvements or additions to documentation
#56
opened Feb 13, 2025 by
hz0ne
Add issue template
documentation
Improvements or additions to documentation
#48
opened Feb 12, 2025 by
Yikun
Abnormal First Token Output on 910B GPU during Inference
bug
Something isn't working
#46
opened Feb 11, 2025 by
Jozenn
Does this project support the deployment of deepseek-v3 and deepseek-r1 on Ascend?
new model
#39
opened Feb 11, 2025 by
Kangzf1996
Question about the difference of inference results between NPU and GPU
#31
opened Feb 11, 2025 by
AIR-hl
Add doc for benchmark and profiling on Ascend NPU
documentation
Improvements or additions to documentation
#26
opened Feb 10, 2025 by
Yikun
Glad to see that vLLM has officially added support for the Ascend hardware backend!
#1
opened Jan 29, 2025 by
liaoyanqing666
ProTip!
Add no:assignee to see everything that’s not assigned.