llama.cpp
7733f0c7
- ggml : support AVX512VNNI (#6280)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
ggml : support AVX512VNNI (#6280) This change causes some quants (e.g. Q4_0, Q8_0) to go faster on some architectures (e.g. AMD Zen 4).
References
#6280 - Support AVX512VNNI
Author
jart
Parents
a32b77c4
Loading