text-generation-inference
2a10a28d
- force attn to flashdecoding
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
286 days ago
force attn to flashdecoding
References
add_chunked_atn
#3162 - Add chunked attn for L4
Author
mht-sharma
Parents
a7353c35
Loading