Releases: NeoZhangJianyu/llama.cpp
Releases · NeoZhangJianyu/llama.cpp
b2466
increase igpu cluster limit (#6159)
b2460
flake.lock: Update Flake lock file updates: • Updated input 'nixpkgs': 'github:NixOS/nixpkgs/9df3e30ce24fd28c7b3e2de0d986769db5d6225d' (2024-03-06) → 'github:NixOS/nixpkgs/d691274a972b3165335d261cc4671335f5c67de9' (2024-03-14)
b2437
fix set main gpu error (#6073)
b2431
llama : fix integer overflow during quantization (#6063)
b2409
ci : remove tidy-review (#6021)
b2408
ggml : reuse quantum structs across backends (#5943) * ggml : reuse quant blocks across backends ggml-ci * ggml : define helper constants only for CUDA and SYCL ggml-ci * ggml : define helper quantum constants for SYCL ggml-ci
b2407
ggml : fix UB in IQ2_S and IQ3_S (#6012)
b2405
grammar : fix unnecessarily retained pointer to rules (#6003)
b2351
compare-llama-bench.py : remove mul_mat_q (#5892)
b2343
fix speculative decoding build on windows (#5874)