Skip to content

Releases: ggerganov/llama.cpp

b4126

18 Nov 18:38
d3481e6
Compare
Choose a tag to compare
cuda : only use native when supported by cmake (#10389)

b4122

18 Nov 11:09
9b75f03
Compare
Choose a tag to compare
Vulkan: Fix device info output format specifiers (#10366)

* Vulkan: Fix device info output format specifiers

* Vulkan: Use zu printf specifier for size_t instead of ld

b4120

17 Nov 23:33
76e9e58
Compare
Choose a tag to compare
CUDA: fix MMV kernel being used for FP16 src1 (#10357)

b4118

17 Nov 12:38
be5cacc
Compare
Choose a tag to compare
llama : only use default buffer types for the KV cache (#10358)

b4115

17 Nov 10:43
a431782
Compare
Choose a tag to compare
ggml : fix undefined reference to 'getcpu' (#10354)

https://github.com/ggerganov/llama.cpp/issues/10352

b4114

17 Nov 09:49
c3ea58a
Compare
Choose a tag to compare
CUDA: remove DMMV, consolidate F16 mult mat vec (#10318)

b4113

17 Nov 09:47
467576b
Compare
Choose a tag to compare
CMake: default to -arch=native for CUDA build (#10320)

b4112

17 Nov 07:55
eda7e1d
Compare
Choose a tag to compare
ggml : fix possible buffer use after free in sched reserve (#9930)

b4111

17 Nov 07:34
Compare
Choose a tag to compare
ggml : inttypes.h -> cinttypes (#0)

ggml-ci

b4103

16 Nov 23:32
4e54be0
Compare
Choose a tag to compare
llama/ex: remove --logdir argument (#10339)