Skip to content

Issues: ggerganov/llama.cpp

changelog : libllama API
#9289 opened Sep 3, 2024 by ggerganov
Open 1
changelog : llama-server REST API
#9291 opened Sep 3, 2024 by ggerganov
Open 7
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

Feature Request: better cross entropy loss CUDA kernel enhancement New feature or request
#10467 opened Nov 23, 2024 by JohannesGaessler
4 tasks done
Support for Macro-o1 by alibaba enhancement New feature or request
#10461 opened Nov 23, 2024 by Meshwa428
4 tasks done
ggml : add ANE backend help wanted Extra attention is needed research 🔬
#10453 opened Nov 22, 2024 by ggerganov
Bug: 【CANN】ggml-cann/aclnn_ops.cpp:3007: GGML_ASSERT(n_dims == src0->ne[0]) failed bug-unconfirmed critical severity Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#10451 opened Nov 22, 2024 by zyp2
Bug: Heavy throttling during token generation on Apple Silicon bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#10444 opened Nov 21, 2024 by Azirine
Bug: Flash Attention performs worse under ROCM bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#10439 opened Nov 20, 2024 by Mushoz
Why server slot's cache_prompt is false by default? bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#10427 opened Nov 20, 2024 by Nekotekina
Bug: SYCL builds >= b4069 fail to allocate SYCL0 buffer bug-unconfirmed critical severity Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#10421 opened Nov 20, 2024 by 0xDEADFED5
Bug: Vulkan vk::DeviceLostError with multithreaded environment bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10420 opened Nov 20, 2024 by ddwkim
Bug: run llama.cpp failed with Vulkan-supported and quantized model in Android Termux bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#10406 opened Nov 19, 2024 by linxhome
Research: bench of the llamacpp research 🔬
#10405 opened Nov 19, 2024 by jumbo-q
1 of 5 tasks
Feature Request: Code Explanation Tutoria enhancement New feature or request
#10399 opened Nov 19, 2024 by Tangzhongyi834
4 tasks done
Bug: Server hangs when number of threads used for decoding > number of CPUs it runs on bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#10397 opened Nov 19, 2024 by KevinRSX
Feature Request: [CANN] Use the RoPE operator provided by aclnn enhancement New feature or request
#10396 opened Nov 19, 2024 by noemotiovon
4 tasks done
Qwen 32B: server breaks stream abruptly when above 9K context bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10393 opened Nov 18, 2024 by JeroenAdam
Research: Design of llama-bench research 🔬
#10386 opened Nov 18, 2024 by jumbo-q
1 of 5 tasks
Bug: flash-attn can't use bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10378 opened Nov 18, 2024 by Tangzhongyi834
Feature Request: Apply LoRA adapters per-request enhancement New feature or request
#10377 opened Nov 18, 2024 by ngxson
4 tasks done
Bug: No docs explain the value for cache-type-k/v bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10373 opened Nov 18, 2024 by phazei
ProTip! Add no:assignee to see everything that’s not assigned.