Releases: teleprint-me/llama.cpp
Releases · teleprint-me/llama.cpp
b3290
Define and optimize RDNA1 (#8085)
b3284
Adding step to `clean` target to remove legacy binary names to reduce…
b3268
flake.lock: Update (#8218)
b3266
llama: Add attention and final logit soft-capping, update scaling fac…
b3264
json: attempt to skip slow tests when running under emulator (#8189)
b3209
ggml : remove ggml_task_type and GGML_PERF (#8017) * ggml : remove ggml_task_type and GGML_PERF * check abort_callback on main thread only * vulkan : remove usage of ggml_compute_params * remove LLAMA_PERF
b3203
Removing extra blank lines that were breaking Lint. (#8067)
b3196
ggml : AVX IQ quants (#7845) * initial iq4_xs * fix ci * iq4_nl * iq1_m * iq1_s * iq2_xxs * iq3_xxs * iq2_s * iq2_xs * iq3_s before sllv * iq3_s * iq3_s small fix * iq3_s sllv can be safely replaced with sse multiply
b3184
ggml : synchronize threads using barriers (#7993)
b3182
[SYCL] refactor (#6408) * seperate lower precision GEMM from the main files * fix workgroup size hardcode