llama.cpp
c4ded1a8
- llama : make pos_bias contiguous for CUDA
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
llama : make pos_bias contiguous for CUDA
References
#8141 - Inference support for T5 and FLAN-T5 model families
Author
sszymczy
Parents
bad0cafe
Loading