Skip to content

Releases: teleprint-me/llama.cpp

b1960

23 Jan 17:20
26d6076
Compare
Choose a tag to compare
metal : disable support for MUL_MAT F32 x F16

b1954

23 Jan 02:05
011e8ec
Compare
Choose a tag to compare
llama : fix not enough space in buffer with Qwen (#5086)

b1893

16 Jan 18:20
bee938d
Compare
Choose a tag to compare
nix: remove nixConfig from flake.nix (#4984)

b1886

16 Jan 17:15
862f5e4
Compare
Choose a tag to compare
android : introduce starter project example (#4926)

* Introduce starter project for Android

Based on examples/llama.swiftui.

* Add github workflow

* Set NDK version

* Only build arm64-v8a in CI

* Sync bench code

* Rename CI prop to skip-armeabi-v7a

* Remove unused tests

b1879

16 Jan 03:57
3e5ca79
Compare
Choose a tag to compare
pass cpu-architecture arguments only to host code (C;C++) (#4943)

b1878

15 Jan 17:03
4483396
Compare
Choose a tag to compare
llama : apply classifier-free guidance to logits directly (#4951)

b1874

15 Jan 07:54
4a3156d
Compare
Choose a tag to compare
CUDA: faster dequantize kernels for Q4_0 and Q4_1 (#4938)

Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>

b1873

14 Jan 16:59
a836c8f
Compare
Choose a tag to compare
llama : fix missing quotes (#4937)

b1863

14 Jan 05:10
76484fb
Compare
Choose a tag to compare
sync : ggml

b1848

12 Jan 22:28
de473f5
Compare
Choose a tag to compare
sync : ggml