text-generation-inference
c9bdaa8b
- feat(server): reduce mlp and attn in one op for flash neox (#145)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
feat(server): reduce mlp and attn in one op for flash neox (#145)
References
#145 - feat(server): reduce mlp and attn in one op for flash neox
Author
OlivierDehaene
Parents
f0000689
Loading