vllm
8fae54fa
- [Linear Attention] fix bug for linear attention + prefix caching + reset_prefix_cache (#35157)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 days ago
[Linear Attention] fix bug for linear attention + prefix caching + reset_prefix_cache (#35157) Signed-off-by: Chen Zhang <zhangch99@outlook.com>
References
#35157 - [Linear Attention] fix bug for linear attention + prefix caching + reset_prefix_cache
Author
heheda12345
Parents
f7967577
Loading