-
Notifications
You must be signed in to change notification settings - Fork 10.2k
Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Eval bug: llama_model_quantize: failed to quantize: unknown model architecture: 'clip'
bug-unconfirmed
#11249
opened Jan 15, 2025 by
fangbaolei
Compile bug: ggml_vulkan:compiling shaders aborted on qualcomm adreno 750 gpu
bug-unconfirmed
#11248
opened Jan 15, 2025 by
AndreaChiChengdu
Eval bug: llama cpp becomes slower as the number of threads -t increases
bug-unconfirmed
#11247
opened Jan 15, 2025 by
wathuta
Feature Request: Support for iFlytek Spark 13B
enhancement
New feature or request
#11232
opened Jan 14, 2025 by
raulbalmez
4 tasks done
Misc. bug: Kompute models fail and struggles where Vulkan works fine
bug-unconfirmed
#11217
opened Jan 13, 2025 by
pepijndevos
"CPU_AARCH64 model buffer" appears when not using AARCH64
bug-unconfirmed
#11204
opened Jan 12, 2025 by
pt13762104
Feature Request: Better chat UX for llama-cli
enhancement
New feature or request
#11202
opened Jan 12, 2025 by
ngxson
Feature Request: Swap Embedding models
enhancement
New feature or request
#11199
opened Jan 11, 2025 by
CHesketh76
4 tasks done
Eval bug: Crash with filesystem error when run while in a directory containing files with certain names
bug
Something isn't working
#11198
opened Jan 11, 2025 by
ScarletEmerald
Misc. bug: Inconsistent responses on subsequent requests with same prompt
bug-unconfirmed
#11197
opened Jan 11, 2025 by
GlasslessPizza
Misc. bug: Docker Image llama-quantize Segmentation fault
bug-unconfirmed
#11196
opened Jan 11, 2025 by
aria3ppp
Feature Request: index.html.gz as a separate distributable
enhancement
New feature or request
#11184
opened Jan 11, 2025 by
charleswg
4 tasks done
Feature request: Up/down arrow to cycle through previous messages
#11182
opened Jan 10, 2025 by
andy144
Eval bug: Add Support for stable-diffusion-3.5-large-turbo-gguf/sd3.5_large_turbo-Q4_0.gguf
bug-unconfirmed
#11181
opened Jan 10, 2025 by
usmandilmeer
Misc. bug: missing chat-template support in llama-run
bug-unconfirmed
#11178
opened Jan 10, 2025 by
engelmi
Misc. bug: empty answer for "long" prompt
bug-unconfirmed
#11176
opened Jan 10, 2025 by
leonardogiacobbe
Feature Request: Way to kill an inference without killing the server
enhancement
New feature or request
#11173
opened Jan 10, 2025 by
paoletto
4 tasks done
Misc. bug: model warmup doesn't work correctly for MoE models
bug-unconfirmed
#11163
opened Jan 9, 2025 by
cpumaxx
Misc. bug: Very bad performance of latest llama.cpp HIP builds with AMD GPU
bug-unconfirmed
#11153
opened Jan 9, 2025 by
http403
Feature Request: Return sliders into llama-server settings interface
enhancement
New feature or request
#11145
opened Jan 8, 2025 by
DaLiV
4 tasks done
Misc. bug: llama-server - shared libraries after build 4409, last working 4406
bug-unconfirmed
#11144
opened Jan 8, 2025 by
DaLiV
server : add support for multiple responses
server/api
server
#11142
opened Jan 8, 2025 by
ggerganov
DeepSeek Models (V2/V3) Hang with ROCm Backend
bug-unconfirmed
#11141
opened Jan 8, 2025 by
emuchogu
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.