text-generation-inference
feat(server): auto max_batch_total_tokens for flash att models
#630
Merged

Loading