-
Notifications
You must be signed in to change notification settings - Fork 11k
Issues: ggml-org/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Misc. bug: llama-rpc crashes when deciding memory on CPU with CUDA_VISIBLE_DEVICES=""
bug-unconfirmed
#12203
opened Mar 5, 2025 by
woof-dog
Eval bug: Granite Vision 3.1 and 3.2 Surgery Script Found 0 Tensors to Extract
bug-unconfirmed
#12202
opened Mar 5, 2025 by
taronaeo
[Metal] Context init optimization opportunity: metal library is compiled for every llama context
#12199
opened Mar 5, 2025 by
iboB
Misc. bug: [Server] Crashes with a coredump during termination
bug-unconfirmed
#12180
opened Mar 4, 2025 by
qnixsynapse
Misc. bug: Denial of Service (crash) when using verbose output with input tokens that are not in printable range.
bug
Something isn't working
good first issue
Good for newcomers
#12178
opened Mar 4, 2025 by
avioligo
Eval bug: Server returns 500 error on /api/generate and /api/chat requests
bug-unconfirmed
#12176
opened Mar 4, 2025 by
blues-alex
Misc. bug: The inference speed of llama-server is one-third of that of llama-cli
bug
Something isn't working
#12171
opened Mar 4, 2025 by
zts9989
Compile bug: issue compiling in ubuntu (desktop and server version) using virtualbox
bug-unconfirmed
#12164
opened Mar 3, 2025 by
sandboxyer
Misc. bug: Calculating the position of kv cache error in llama sever
bug-unconfirmed
#12160
opened Mar 3, 2025 by
Clauszy
Eval bug: The answers have some problems with the example/llama.android
bug-unconfirmed
#12158
opened Mar 3, 2025 by
chtfrank
CUDA: HIP: maintain_cuda_graph use of cudaGraphKernelNodeGetParams is incorrect.
#12152
opened Mar 2, 2025 by
IMbackK
Misc. bug: gguf-dump 'newbyteorder' was removed
bug-unconfirmed
#12146
opened Mar 2, 2025 by
dlippold
Feature Request: Implement Qwen2Model
enhancement
New feature or request
#12142
opened Mar 2, 2025 by
wqerrewetw
4 tasks done
Feature Request: Enable cuda 11.4 and cuda arch 3.7
enhancement
New feature or request
#12140
opened Mar 2, 2025 by
ChunkyPanda03
4 tasks done
Feature Request: Proposing User-Customizable RAG Integration in llama.cpp: A Path to Enhanced Contextual Retrieval
enhancement
New feature or request
#12129
opened Mar 1, 2025 by
gnusupport
4 tasks done
Feature Request:
enhancement
New feature or request
#12128
opened Mar 1, 2025 by
gnusupport
4 tasks done
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.