-
Notifications
You must be signed in to change notification settings - Fork 10.5k
Pull requests: ggerganov/llama.cpp
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
HIP: force max threads per block to be 1024
ggml
changes relating to the ggml tensor library for machine learning
#11621
opened Feb 3, 2025 by
fxzjshm
Loading…
server : add try..catch to places not covered by set_exception_handler
examples
server
#11620
opened Feb 3, 2025 by
ngxson
Loading…
HIP: add doc on small default launch bounds
documentation
Improvements or additions to documentation
#11619
opened Feb 3, 2025 by
fxzjshm
Loading…
ci: add bash script to check if llama-impl.h was included erroneously
devops
improvements to build systems and github actions
script
Script related
#11617
opened Feb 3, 2025 by
mofosyne
Loading…
Clean up Test Script + Update it to work on Instruct Tuned Models
examples
SYCL
https://en.wikipedia.org/wiki/SYCL - GPU programming language
#11610
opened Feb 3, 2025 by
Mr-Thack
Loading…
scripts: added inline script metadata per PEP 723
python
python script changes
script
Script related
#11597
opened Feb 2, 2025 by
isaac-mcfadyen
Loading…
De-duplicate fmt and format functions and optimize
examples
#11596
opened Feb 2, 2025 by
ericcurtin
Loading…
vulkan: add specific MMV kernels for IQ2 and IQ3 quants + optimizations
devops
improvements to build systems and github actions
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#11595
opened Feb 2, 2025 by
remyoudompheng
•
Draft
vulkan: add environment variable to avoid VRAM allocation
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#11592
opened Feb 2, 2025 by
wbruna
Loading…
NUMA-aware KV cache buffer type (experimental)
ggml
changes relating to the ggml tensor library for machine learning
#11580
opened Feb 1, 2025 by
fairydreaming
•
Draft
Add support for Deepseek-R1 flash attention
ggml
changes relating to the ggml tensor library for machine learning
Nvidia GPU
Issues specific to Nvidia GPUs
#11557
opened Jan 31, 2025 by
siddartha-RE
Loading…
tool-call: add support to llama-cli using new --tools arg
examples
#11556
opened Jan 31, 2025 by
bandoti
Loading…
vulkan: use smaller combined allocations to avoid fragmentation
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#11551
opened Jan 31, 2025 by
jeffbolznv
Loading…
readme : add UI of ChatPDFLocal which is based by llama.cpp for running LLMs on Mac
#11534
opened Jan 31, 2025 by
ljeagle
Loading…
vulkan: initial support for IQ1_S and IQ1_M quantizations
devops
improvements to build systems and github actions
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#11528
opened Jan 30, 2025 by
remyoudompheng
•
Draft
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.