llama.cpp
0e6ff004
- CUDA: larger SRAM reads for tile FA, AMD FP16 dot (#15927)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
8 days ago
CUDA: larger SRAM reads for tile FA, AMD FP16 dot (#15927) * CUDA: larger SRAM reads for tile FA, AMD FP16 dot * fix logic for availability of v_dot2_f32_f16
References
#15927 - CUDA: larger SRAM reads for tile FA, AMD FP16 dot
Author
JohannesGaessler
Parents
df082f56
Loading