llama.cpp
Add support for QRWKV6 hybrid models & slight optimization for RWKV6
#11001
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
12
Changes
View On
GitHub
Previous Change (CTRL+↑)
Next Change (CTRL+↓)
Expand Context Lines
Collapse Context Lines
Hide Minimap (CTRL+M)
Files
23
Threads
README.md
convert_hf_to_gguf.py
ggml
include
ggml.h
src
ggml-cpu
ggml-cpu.c
ggml-cuda
ggml-cuda.cu
gla.cu
gla.cuh
wkv6.cu
ggml-sycl
wkv6.cpp
ggml-vulkan
ggml-vulkan.cpp
ggml.c
gguf-py/gguf
constants.py
gguf_writer.py
tensor_mapping.py
src
llama-arch.cpp
llama-arch.h
llama-hparams.cpp
llama-hparams.h
llama-model.cpp
llama-model.h
llama-quant.cpp
llama.cpp
tests
test-backend-ops.cpp
Loading comments...
Loading