text-generation-inference
feat(server): auto max_batch_total_tokens for flash att models
#630
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
19
Changes
View On
GitHub
Loading