llama.cpp
0e6ff004 - CUDA: larger SRAM reads for tile FA, AMD FP16 dot (#15927)

Commit
8 days ago
CUDA: larger SRAM reads for tile FA, AMD FP16 dot (#15927) * CUDA: larger SRAM reads for tile FA, AMD FP16 dot * fix logic for availability of v_dot2_f32_f16
Parents
Loading