KEMBAR78
Release b6109 · ggml-org/llama.cpp · GitHub
Skip to content

b6109

Choose a tag to compare

@github-actions github-actions released this 07 Aug 10:00
1d72c84
CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16 (#15131)

* CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16