text-generation-inference
fe80f536
- feat(server): auto max_batch_total_tokens for flash att models (#630)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
feat(server): auto max_batch_total_tokens for flash att models (#630)
References
#630 - feat(server): auto max_batch_total_tokens for flash att models
Author
OlivierDehaene
Parents
5e6ddfd6
Loading