Releases: ggml-org/llama.cpp
Releases · ggml-org/llama.cpp
b5794
vulkan: Split large mul_mat_id to fit in shared memory (#14451)
b5793
add GELU_ERF (#14455)
b5792
ggml : remove trailing whitespace (#0)
b5788
opencl : add GEGLU, REGLU, SWIGLU (#14456)
b5787
Add Conv2d for CPU (#14388) * Conv2D: Add CPU version * Half decent * Tiled approach for F32 * remove file * Fix tests * Support F16 operations * add assert about size * Review: further formatting fixes, add assert and use CPU version of fp32->fp16
b5785
metal : disable fast-math for some cpy kernels (#14460) * metal : disable fast-math for some cpy kernels ggml-ci * cont : disable for q4_1 ggml-ci * cont : disable for iq4_nl ggml-ci
b5784
ggml-cpu: sycl: Re-enable exp f16 (#14462)
b5783
test-backend-ops : disable llama test (#14461)
b5782
cmake : Remove redundant include path in CMakeLists.txt (#14452) * Update docker.yml 修改docker.yml文件中的内容使其停止周期性的运行该workflow,如果想要运行该workflow可以手动启动 * Remove redundant include path in CMakeLists.txt The parent directory '..' was removed from the include directories for the ggml-cpu-feats target, to avoid unnecessary include paths. * Enable scheduled Docker image builds Uncomments the workflow schedule to trigger daily Docker image rebuilds at 04:12 UTC, improving automation and keeping images up to date.
b5780
server : support jinja extra template kwargs (Qwen3 enable_thinking f…