llama.cpp
8d8d54f8
- ggml : skip nops in compute_forward
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
ggml : skip nops in compute_forward
References
#3749 - cuda : add batched cuBLAS GEMM for faster attention
Author
ggerganov
Parents
6a30bf3e
Loading