text-generation-inference
e74bd41e
- feat(server): add paged attention to flash models (#516)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
feat(server): add paged attention to flash models (#516) Closes #478
Author
OlivierDehaene
Parents
70f485bf
Loading