Releases: teleprint-me/llama.cpp
Releases · teleprint-me/llama.cpp
b1960
metal : disable support for MUL_MAT F32 x F16
b1954
llama : fix not enough space in buffer with Qwen (#5086)
b1893
nix: remove nixConfig from flake.nix (#4984)
b1886
android : introduce starter project example (#4926) * Introduce starter project for Android Based on examples/llama.swiftui. * Add github workflow * Set NDK version * Only build arm64-v8a in CI * Sync bench code * Rename CI prop to skip-armeabi-v7a * Remove unused tests
b1879
pass cpu-architecture arguments only to host code (C;C++) (#4943)
b1878
llama : apply classifier-free guidance to logits directly (#4951)
b1874
CUDA: faster dequantize kernels for Q4_0 and Q4_1 (#4938) Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
b1873
llama : fix missing quotes (#4937)
b1863
sync : ggml
b1848
sync : ggml