llama.cpp
8f900abf - CUDA: faster softmax via shared memory + fp16 math (#4742)

Commit
2 years ago
CUDA: faster softmax via shared memory + fp16 math (#4742)
Parents
Loading