llama.cpp
CUDA: optimize MMQ int8 tensor core performance
#8062
Merged

Loading