llama.cpp
a743d76a
- CUDA: generalize FP16 fattn vec kernel (#7061)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
CUDA: generalize FP16 fattn vec kernel (#7061) * CUDA: generalize FP16 fattn vec kernel * disable unsupported head sizes for AMD in test * try AMD fix * fix batch size 2-8 * partially revert changes
References
#7061 - CUDA: generalize FP16 fattn vec kernel
Author
JohannesGaessler
Parents
f31ec120
Loading