vllm
1173804d
- [Bugfix] Fix TP inference for Flex attention backend (#19657)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
199 days ago
[Bugfix] Fix TP inference for Flex attention backend (#19657) Signed-off-by: Isotr0py <2037008807@qq.com>
References
#19657 - [Bugfix] Fix TP inference for Flex attention backend
Author
Isotr0py
Parents
4d542402
Loading