text-generation-inference
22fb1be5 - Fix cache block size for flash decoding (#2351)

Commit
1 year ago
Fix cache block size for flash decoding (#2351) * Fix cache block size for flash decoding This seems to have been accidentally dropped during the TRT-LLM PR rebase. * Also run CI on changes to `backends`
Author
Parents
Loading