forked from ggml-org/llama.cpp
-
Notifications
You must be signed in to change notification settings - Fork 29
Pull requests: tetherto/qvac-fabric-llm.cpp
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Rebase on latest llama.cpp
android
Apple Metal
Ascend NPU
build
devops
documentation
Improvements or additions to documentation
examples
ggml
Hexagon
IBM zDNN
jinja parser
model
nix
Nvidia GPU
OpenCL
python
script
server
SYCL
testing
Vulkan
WebGPU
#116
opened Mar 29, 2026 by
zoq
Loading…
TurboQuant: KV cache quantization with Hadamard transform (TBQ3_0 / TBQ4_0)
ggml
python
script
testing
Vulkan
#115
opened Mar 27, 2026 by
jesusmb1995
•
Draft
15 of 23 tasks
Revert to C style construciton for termux portability
ggml
Vulkan
#114
opened Mar 27, 2026 by
arsalan-anwari-tether
Loading…
cuda: restrict out_prod support to f32 inputs
ggml
Nvidia GPU
#110
opened Mar 19, 2026 by
GuthL
Loading…
ProTip!
Adding no:label will show everything without a label.