Releases: ggml-org/llama.cpp
Releases · ggml-org/llama.cpp
b7050
sched : fix reserve ignoring user tensor assignments (#17232)
b7049
ggml-cpu : add RISC-V vector intrinsic support for silu and cvar oper…
b7048
metal: accelerated conv2d (#17175) * metal: accelerated conv2d * cont : cleanup --------- Co-authored-by: bghira <[email protected]> Co-authored-by: Georgi Gerganov <[email protected]>
b7047
Revert "ggml-cpu: handle 3d tensors in repack mat_mul (#17030)" (#17233) This reverts commit 1c398dc9eca9c366ce98deb0e6f3538e444ebc8a.
b7046
ggml-cpu : use template for argsort (#17222)
b7045
CANN: Add cross_entropy_loss op support (#16886) * update L2_NORM op support * update L2_NORM op support * remove extra whitespace * cann: update cross_entropy_loss op support * remove trailing whitespaces * rebase the latest code in the main repository and remove the l2_norm operator that already exists in another pull request. * undo the l2_norm operator deletion
b7044
CUDA: fuse rope + set_rows (#16884) * CUDA: add fused rope * move k forward_expand up * create helper function instead of re-using params * make assert statement more in line with comment * rope_norm: coalesced writes to global mem
b7042
vocab : correct bounds check for UGM XCDA array access (#17215)
b7041
CUDA: static assert to prevent misuse of memcpy_1 (#17198)
b7039
ggml : use std::sort in ggml_argsort CPU implementation (#17211) * ggml : use std::sort in ggml_argsort CPU implementation * cont : add missing header