llama.cpp
51f0bd50
- Remove custom pre attention scaling and use computed value instead.
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
Remove custom pre attention scaling and use computed value instead.
References
add-gemma2-soft-capping
#8197 - Add attention and final logit soft-capping, update scaling factor to Gemma2
Author
abetlen
Parents
a8942790
Loading